Skip to content

v0.2.0 - 2025-04-22

🚀 Features

  • (callbacks) Enhanced callbacks for flexibility.
  • (force) Added force flag to save methods for file overwrites.
  • (ppo) Added LiquidPPO algorithm.
  • (handler) Added file saving for training completion details.
  • (ncp) Added multiple weight initialization options.
  • (sac) Added LiquidSAC agent for continuous action spaces.
  • (sac) Added LiquidSACDiscrete agent for discrete action spaces.
  • (neuroflow) Added main logic for NeuroFlow.
  • (agent) Added NeuroFlowDiscrete agent.

🐛 Bug Fixes

  • (ddpg) Fixed noise handling and prediction bugs.
  • (cell) Fixed sparsity_mask assignment bug.
  • (params) Fixed parameter counts in in DDPG.
  • (ppo) Fixed PPO callback bugs and metric tracking.
  • (config) Fixed bug with train_params in RLAgentConfig.
  • (load) Fixed model loading bug.
  • (buffer) Fixed warm method bug when num_envs=1.
  • (buffer) Fixed save bug where directories don't exist.

💼 Other

  • (box) Added Gymnasium box2d environments by default.

🚜 Refactor

  • (ncp) Added update_mask helper methods.
  • (metrics) Updated training metrics name for clarity.
  • (metrics) Simplified metric classes using base class.
  • (train) Refactored TrainHandler, TrainConfig to simplify.
  • (buffer) Added Actor hidden state to buffer.
  • (seed) Improved random seed generation.
  • (save) Simplified save, load method implementations.
  • (sac) Moved SAC agents to separate folder for simplicity.
  • (ncp) Renamed NCPModule -> LiquidNCPModule for clarity.
  • (agents) Refactored framework to centre around NeuroFlow.
  • (save) Moved completed.json to save directory.
  • (warm) Improved buffer warming step implementation.
  • (utils) Simplified capture utility methods.