Skip to content

Latest commit

 

History

History
101 lines (59 loc) · 3.62 KB

NEWS.md

File metadata and controls

101 lines (59 loc) · 3.62 KB

ReinforcementLearning.jl Release Notes

ReinforcementLearningEnvironments.jl

ReinforcementLearningCore.jl

ReinforcementLearningZoo.jl

v0.5.0

  • Update the complete SAC implementation and modify some details based on the original paper. #365
  • Add some extra keyword parameters for BehaviorCloningPolicy to use it online. #390

ReinforcementLearningDatasets.jl

v0.1.0

  • Add functionality for fetching d4rl datasets as an iterable DataSet. Credits: https://arxiv.org/abs/2004.07219
  • This supports d4rl and d4rl-pybullet and Google Research DQN atari datasets.
  • Uses DataDeps for data dependency management.
  • This package also supports RL Unplugged Datasets.
  • Support for google-research/deep_ope added.

ReinforcementLearningBase.jl

v0.9.6

  • Implement Base.:(==) for Space. #428

v0.9.5

  • Add default Base.:(==) and Base.hash method for AbstractEnv. #348

ReinforcementLearningCore.jl

v0.8.3

  • Add extra two optional keyword arguments (min_σ and max_σ) in GaussianNetwork to clip the output of logσ. #428

v0.8.2

  • Add GaussianNetwork and DuelingNetwork into ReinforcementLearningCore.jl as general components. #370
  • Export WeightedSoftmaxExplorer. #382

v0.8.1

  • Minor bug & typo fixes

v0.8.0

  • Removed ResizeImage preprocessor to reduce the dependency of ImageTransformations.
  • Show unicode plot at the end of an experiment in the TotalRewardPerEpisode hook.

ReinforcementLearningZoo.jl

v0.4.1

  • Make keyword argument n_actions in TabularPolicy optional. #300

v0.4.0

  • Moved all the experiments into a new package ReinforcementLearningExperiments.jl. The related dependencies are also removed (BSON.jl, StableRNGs.jl, TensorBoardLogger.jl).

ReinforcementLearningEnvironments.jl

v0.6.4-dev

  • Add GraphShortestPathEnv. #445

v0.6.3

v0.6.2

  • Add SequentialEnv environment wrapper to turn a simultaneous environment into a sequential one.

v0.6.1

  • Drop GR in RLEnvs and lazily load ploting functions.#309, #310

v0.6.0

  • Set AcrobotEnv into lazy loading to reduce the dependency of OrdinaryDiffEq.

ReinforcementLearningExperiments.jl

v0.1.0