Skip to content

Releases: JuliaReinforcementLearning/ReinforcementLearning.jl

v0.10.2

03 Mar 11:54
d1971f5
Compare
Choose a tag to compare

ReinforcementLearning v0.10.2

Diff since v0.10.1

Closed issues:

  • Add procgen (#126)
  • CI fails with [email protected] (#572)
  • Missing docs for TDLearner (#580)
  • Add an environment wrapper to IsaacGym (#619)
  • How to run this source code in vscode? (#623)
  • Examples of multidimensional continous actions (#676)
  • Base.copy not implemented for the TicTacToe environment (#678)
  • Broken link to src (#693)
  • Support Brax (#696)
  • PPO on environments with multiple action dimensions? (#703)
  • Can't checkout RLCore for development (#704)
  • Setup sponsor related info (#730)
  • new _run() (#731)
  • PPOPolicy training: ERROR: DomainError with NaN: Normal: the condition σ >= zero(σ) is not satisfied. (#739)
  • Code Readability (#740)
  • MultiThreadEnv not available in ReinforcementLearningZoo (#741)
  • ReinforcementLearningExperiment dependencies fail to precompile (#744)
  • tanh normalization destabilizes learning with GaussianNetwork (#745)
  • Custom Environment Passes RLBase.test_runnable!(env) but infinite hangs and crashes when run. (#757)
  • Collect both number of steps and rewards in a single hook (#763)
  • Every single environment / experiment crashes with following error: (#766)
  • Neural Network Approximator based policies not working (#770)
  • "params not defined," "JuliaRL_BasicDQN_CartPole" (#778)

Merged pull requests:

  • WIP: Add MPO in zoo (#604) (@HenriDeh)
  • Episode reset condition (#621) (@HenriDeh)
  • Add a categorical Network (#625) (@HenriDeh)
  • Use Trajectories.jl instead (#632) (@findmyway)
  • added basic doc for TDLearner (#649) (@baedan)
  • Add JuliaRL_DQN_CartPole (#650) (@findmyway)
  • enable OpenSpiel (#691) (@findmyway)
  • Small improvements for TicTacToeEnv (#692) (@jonathan-laurent)
  • Update the "how to implement a new algorithm" (#695) (@HenriDeh)
  • Fix typo in algorithm implementation docs (#697) (@mplemay)
  • add PrioritizedDQN (#698) (@findmyway)
  • add QRDQN (#699) (@findmyway)
  • add REMDQN (#708) (@findmyway)
  • add IQN (#710) (@findmyway)
  • checkin Mainifest.toml (#711) (@findmyway)
  • CompatHelper: bump compat for "ReinforcementLearningCore" to "0.8" (#712) (@github-actions[bot])
  • CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.6" (#713) (@github-actions[bot])
  • CompatHelper: bump compat for "ReinforcementLearningZoo" to "0.5" (#714) (@github-actions[bot])
  • CompatHelper: bump compat for "AbstractTrees" to "0.4" for package ReinforcementLearningBase (#715) (@github-actions[bot])
  • CompatHelper: bump compat for "Functors" to "0.3" for package ReinforcementLearningCore (#717) (@github-actions[bot])
  • CompatHelper: bump compat for "UnicodePlots" to "3" for package ReinforcementLearningCore (#718) (@github-actions[bot])
  • CompatHelper: bump compat for "ReinforcementLearningCore" to "0.8" for package ReinforcementLearningZoo (#720) (@github-actions[bot])
  • CompatHelper: bump compat for "Functors" to "0.3" for package ReinforcementLearningZoo (#721) (@github-actions[bot])
  • CompatHelper: add new compat entry for "StableRNGs" at version "1" for package ReinforcementLearningExperiments (#722) (@github-actions[bot])
  • CompatHelper: bump compat for "ReinforcementLearning" to "0.10" for package ReinforcementLearningExperiments (#723) (@github-actions[bot])
  • add rainbow (#724) (@findmyway)
  • Adapted SAC to support MultiThreadedEnv (#726) (@BigFood2307)
  • Add the number of episodes (#727) (@ll7)
  • docs: add ll7 as a contributor for doc (#728) (@allcontributors[bot])
  • Add struct view (#732) (@findmyway)
  • add VPG (#733) (@findmyway)
  • CompatHelper: add new compat entry for "Distributions" at version "0.25" for package ReinforcementLearningZoo (#734) (@github-actions[bot])
  • CompatHelper: add new compat entry for "Distributions" at version "0.25" for package ReinforcementLearningExperiments (#735) (@github-actions[bot])
  • fixed hyperlink in readme (#742) (@mplemay)
  • docs: add mplemay as a contributor for doc (#743) (@allcontributors[bot])
  • Create FUNDING.yml (#746) (@findmyway)
  • TRPO (#747) (@baedan)
  • CompatHelper: bump compat for "CommonRLSpaces" to "0.2" for package ReinforcementLearningBase (#748) (@github-actions[bot])
  • Fix parameter names for AsyncTrajectoryStyle (#749) (@ludvigk)
  • Update DoEveryNEpisode hook to new api (#750) (@ludvigk)
  • docs: add ludvigk as a contributor for code (#751) (@allcontributors[bot])
  • Update TwinNetwork (#752) (@ludvigk)
  • Typo in hooks docs (#754) (@kir0ul)
  • CommonRLSpace -> DomainSets (#756) (@findmyway)
  • Fix typo (#767) (@jeremiahpslewis)
  • Fix typo (#768) (@jeremiahpslewis)
  • Fix TD Learner so that it handles MultiAgent/Simultaneous with NoOp (#769) (@jeremiahpslewis)
  • Bump RLBase compat to 0.11 (#771) (@HenriDeh)
  • Remove manifest from the repo (#773) (@HenriDeh)
  • import params and gradient (#774) (@HenriDeh)
  • fix compat (#775) (@HenriDeh)
  • Trying to reimplement experiments (#776) (@HenriDeh)
  • Add a developer mode (#777) (@HenriDeh)
  • added pettingzoo and one single agent example (#782) (@Mytolo)
  • Update mpo.jl (#783) (@HenriDeh)
  • Reduce unnecessary array allocations (#785) (@jeremiahpslewis)
  • Temporarily disable failing experiment so project tests pass (#787) (@jeremiahpslewis)
  • Fix spellcheck errors (#788) (@jeremiahpslewis)
  • Bug fixes and dependency bump (#789) (@jeremiahpslewis)
  • Pin ReinforcementLearning.jl to pre-refactor versions (#793) (@jeremiahpslewis)

v0.10.1

04 Jun 16:10
2e1de3e
Compare
Choose a tag to compare

ReinforcementLearning v0.10.1

Diff since v0.10.0

Closed issues:

  • Support compression? (#102)
  • State monitoring and fault tolerance (#101)
  • Add card game environments (#196)
  • Add Discrete Batch-Constrained Deep Q-learning (#226)
  • Add a dedicated multi-dimensional space type (#268)
  • PyCall.getindex in module ReinforcementLearningEnvironments conflict warning (#527)
  • device method definition overwritten (#530)
  • StackFrames bug? (#551)
  • Small performance improvement (#558)
  • Infinite-recursion bug in function is_discrete_space when an object of type ClosedInterval is passed (#566)
  • action_space not defined in tutorial (#569)
  • Warning while precompiling RLCore due to kwargs (#575)
  • Strange Bug with examples CartPoleEnv and RLBase.test_runnable!(RandomWalk1D) (#579)
  • Difficulty Creating a Custom Environment (#581)
  • Missing docs for how to implement a new algorithm (#582)
  • Donation (#595)
  • MultiThreadEnv with custom (continuous) action spaces fails (#596)
  • PPOCartPole fails, source of error included (#605)
  • Bug: Issue with TD3 for multi-dimensional action spaces (#624)
  • ActionTransformedEnv doesn't transform legal_action_space_mask (#642)
  • Bug: Previous example from RLZoo now has a bug (#643)

Merged pull requests:

v0.10.0

08 Oct 12:20
a971df7
Compare
Choose a tag to compare

ReinforcementLearning v0.10.0

Diff since v0.9.0

Closed issues:

  • In DDPG: Add support for vector actions (#138)
  • Add experiments based on offline RL data (#141)
  • Train policy with GymEnv (#175)
  • SARTTrajectory for SAC (#182)
  • PPO related algorithms are broken (#194)
  • ERROR: type RandomPolicy has no field policy (#208)
  • "Getting Started" too long imo (#210)
  • Documentation of environment; actions seems not work. (#222)
  • Documentation of "How to use Tensorboard?": with_logger not defined (#223)
  • Getting figure object; How to get an animation using GR.plot in CartPolEnv (#246)
  • The components of Rainbow (#229)
  • code in get_started seems to be broken (#233)
  • Document how to save/load parameters (#238)
  • Workflow of saving trajectory data (#239)
  • [Call for Contributors] Summer 2021 of Open Source Promotion Plan (#242)
  • Next Release Plan (v0.9) (#247)
  • Add ReinforcementLearningDatasets (#253)
  • Lack of reproducibility of QRDQN CartPole Experiment. (#281)
  • StopAfterNoImprovement hook test fails occasionally (#297)
  • Get error when using ReinforcementLearning (#298)
  • Problems with PGFPlotsX during the install (#301)
  • Plotting CartPole environment in Jupyter (#306)
  • Local development environment setup tips causing error (#320)
  • Question about PER (#328)
  • Docs error in code output (#332)
  • Setup a CI for typo (#336)
  • double code & dysfunctional master branch when downloading package (#341)
  • Precompilation error; using Plots makes a conflict (#349)
  • Problem with running initial tutorial. Using TabularPolicy() generates an UndefinedKeyword error for n_action (#354)
  • Question: Clarification on the RL plots generated by the run() function (#357)
  • prob question for QBasedPolicy (#360)
  • Can evaluate function be used as a component of RLcore? (#369)
  • problem about precompiling the forked package (#377)
  • Question: Can we use packages like DifferentialEquations.jl to evolve or model the environment in ReinforcementLearning.jl (#378)
  • MultiAgentManager does not select correct action space for RockPaperScissorsEnv (#393)
  • Add ReinforcementLearningDatasets.jl (#397)
  • error: dimension mismatch "cannot broadcast array to have fewer dimensions" (#400)
  • SAC policy problems? (#410)
  • Add pre-training hook (#411)
  • Dead links in documentation (#418)
  • Links of show nbview badges in RLExperiments are incorrect (#421)
  • Problem accessing public google cloud storage bucket for RLDatasets.jl (#424)
  • Function to access base env through multiple wrapper layers (#425)
  • The problem of using GaussianNetwork in gpu (#455)
  • Next Release Plan (v0.10) (#460)
  • Error in experiment "JuliaRL_DDPG_Pendulum" (#471)
  • In Windows, ReinforcementLearningDataset.jl encounter a bug (#485)
  • Conditional Testing (#493)
  • Inconsistency of the EpsilonGreedyExplorer selection function (#520)

Merged pull requests:

Read more

v0.9.0

16 May 05:47
98f5f20
Compare
Choose a tag to compare

ReinforcementLearning v0.9.0

Diff since v0.8.0

Closed issues:

  • Roadmap of v0.9 (#65)
  • Classic environments in separate package? (#123)
  • Add dueling DQN (#137)
  • How should ReinforcementLearning.jl be cited ? (#80)
  • Alternative handling of max steps in environment (#140)
  • Add Highway env (#120)
  • Add experiments with GymEnv (#147)
  • Replace Travis with github actions (#151)
  • AbstractStage docstring doesn't render correctly in docs. (#160)
  • List of contributors (#161)
  • Return experiment instead of hook only (#173)
  • Training mode and testing mode (#179)
  • AbstractEnv (#181)
  • define environment of FULL_ACTION_SET (#184)
  • CircularArraySLARTTrajectory instance is not of type CircularArraySLARTTrajectory (#185)
  • Is hook the same thing as "callback"? (#190)
  • Use @threads instead of @sync + @Spawn in MultiThreadEnv? (#191)
  • Blog custom env link typo (#192)
  • Separate envs from algos in Zoo? (#197)
  • Why "examples"? (#198)
  • WandB integration? (#201)
  • Add default implementations for AbstractEnvWrapper (#202)
  • Add configuration in DQNLearner to enable double-dqn by default (#205)
  • Why split repos? (#209)
  • PreActStage clarification (#212)
  • What's a "trace"? (#213)
  • Continuous time supported? (#215)
  • Docs looks ugly in dark mode (#217)
  • Julia 1.6.0 dependency problem with ReinforcementLearningBase/RLBase (#221)
  • Docstring of DoEveryNStep (#225)
  • Update dependency to [email protected] and resolve type piracy of findmax (#227)
  • IQN is broken with [email protected] (#228)
  • Source links in documentation directs to julia repo (#231)
  • PPO strange behaviour from having actions as one element arrays instead of scalar (#234)
  • SAC and GaussianNetwork (#236)
  • Precompilation prohibitively long (#237)
  • An explanation of "how to train policy (agent)" such as Basic_DQN would be valuable (#240)
  • How to guarantee the environment's reproducibility? (#241)
  • Cannot use RLBase.action_space etc. when writing my own environment (#244)
  • ReinforcementLearningZoo.jl experiments (#245)
  • How about making this package compatible with DifferentialEquations.jl? (#249)
  • PPO and multi dimensional actions spaces (#251)
  • Incompatibility with CSVFiles.jl (#256)
  • [RLEnvs] easy access of the length of an action vector (dimension of action space) (#257)
  • Cannot add LinearAlgebra (#258)
  • What's the checkpoints? (#261)
  • PPO is broken when using CUDA (#280)
  • Reinforcement Learning.jl in a RTS (#291)

Merged pull requests:

v0.8.0

26 Jan 00:58
d71899c
Compare
Choose a tag to compare

ReinforcementLearning v0.8.0

Diff since v0.7.0

Closed issues:

  • Document basic environments (#129)
  • Improve interfaces for model exploration and hyperparameter optimization (#28)
  • Support SEED RL (SCALABLE AND EFFICIENT DEEP-RL ) (#62)
  • Rename AbstractAgent to AbstractPolicy (#111)
  • Add a stop condition to terminate the experiment after reaching reward threashold (#112)
  • ACME RL lib by deepmind (#85)
  • Definition of a policy (#86)
  • Add remote trajectories (#87)
  • Base.convert method for DiscreteSpace (#104)
  • Action Space Meaning (#88)
  • Base.in method for EmptySpace (#105)
  • Renaming get_terminal to isterminated (#106)
  • Requesting more informative field names for SharedTrajectory (#113)
  • Suggestion: More informative name for FullActionSet & MinimalActionSet (#107)
  • Returning an AbstractSpace object using get_actions (#108)
  • Split experiments into separate files (#145)
  • Add project.toml for tests (#146)
  • Docs build error (#91)
  • Split out Trajectory & CircularArrayBuffer as independent packages (#114)
  • Requesting explanation for better performance at ... (#115)
  • Add an extra mode when evaluating agent (#116)
  • Why are wrapper environments a part of RLBase instead of RLCore (say)? (#109)
  • The names of keyword arguments in Trajectory is kind of misunderstanding (#117)
  • Check compatibility between agent and environments (#118)
  • Behaviour for hooks for RewardOverridenEnv (#119)
  • StopAfterEpisode with custom DQNL errors beyond a particular Episode Count (#96)
  • ERROR: UndefVarError: NNlib not defined while loading agent (#110)
  • Use JLSO for (de)serialization? (#97)
  • Setup github actions (#98)
  • Fails to load trajectory (#150)
  • Test error in ReinforcementLearningEnvironments.jl (#152)
  • Move preallocations in MultiThreadEnv from global to local (#153)
  • remove @views (#155)
  • error in save & load ElasticCompactSARTSATrajectory (#156)
  • add early stopping in src\core\stop_conditions.jl (#157)
  • add time stamp in load & save function, in file src\components\agents\agent.jl (#158)
  • policies in GPU can not be saved || loaded (#159)
  • code formatting (#165)
  • Purpose of CommonRLInterface (#166)
  • Moving example environments from RLBase to RLEnvs? (#167)
  • Keeping prefix get_ in method names like get_reward (#168)
  • Currently getting an ambiguous method error in ReinforcementLearningCore v0.5.1 (#171)
  • TD3 Implementation (#174)
  • Travis CI Credits (#178)
  • Unrecognized symbols (#180)

Merged pull requests:

v0.7.0

24 Oct 01:01
3a632ca
Compare
Choose a tag to compare

ReinforcementLearning v0.7.0

Diff since v0.6.0

Closed issues:

  • How to define a new environment? (#64)
  • Question about AbstractEnv API (#68)
  • Compatibility issue in ReinforcementLearning & Flux (#74)
  • ERROR: KeyError: key "ArnoldiMethod" not found (#79)
  • I get NNlib error when trying to load a model (#82)
  • "convert" warning (#83)
  • Seg fault on macbook pro (#84)

v0.6.0

06 Aug 00:38
f541772
Compare
Choose a tag to compare

ReinforcementLearning v0.6.0

Diff since v0.5.0

Make environments transparent to agents/policies.
Mooncake release may be delayed.

Closed issues:

  • Failed to precompile ReinforcementLearning (#71)
  • depends on HDF5? (#72)
  • warning and error (#73)

Merged pull requests:

v0.5.0

21 Jun 00:26
913fcb6
Compare
Choose a tag to compare

ReinforcementLearning v0.5.0

Diff since v0.3.0

Closed issues:

  • Box2D environment (#2)
  • bullet3 environment (#7)
  • loadenvironment error (#19)
  • Support alternative deep learning libraries (#20)
  • Random Thoughts on v0.3.0 (#24)
  • Prioritized DQN (#27)
  • A2C (#32)
  • Add built-in support for TensorBoard (#35)
  • Add checkpoints (#36)
  • Improve code coverage (#40)
  • AbstractActionSelector not exported (#41)
  • Params empty - no tracking (#43)
  • Add reproducible examples for Atari environments (#44)
  • StopAfterEpisode with progress meter (#51)
  • Support julia 1.4 (#63)

Merged pull requests:

Preview of v0.4.0

17 Feb 00:28
Compare
Choose a tag to compare
Preview of v0.4.0 Pre-release
Pre-release
v0.4.0-beta

Install TagBot as a GitHub Action (#53)

Preview of v0.4.0

25 Sep 08:05
574bd41
Compare
Choose a tag to compare
Preview of v0.4.0 Pre-release
Pre-release

This is a preview of v0.4.0.

What we have for now:

  • Tabular methods are well tested.
  • Some simple value based methods are implemented.

What are missing:

  • GPU support
  • Some other DRL methods proposed in the issues with tag v0.4.0