Releases: JuliaReinforcementLearning/ReinforcementLearning.jl
Releases · JuliaReinforcementLearning/ReinforcementLearning.jl
v0.10.2
ReinforcementLearning v0.10.2
Closed issues:
- Add procgen (#126)
- CI fails with [email protected] (#572)
- Missing docs for
TDLearner
(#580) - Add an environment wrapper to IsaacGym (#619)
- How to run this source code in vscode? (#623)
- Examples of multidimensional continous actions (#676)
- Base.copy not implemented for the TicTacToe environment (#678)
- Broken link to src (#693)
- Support Brax (#696)
- PPO on environments with multiple action dimensions? (#703)
- Can't checkout RLCore for development (#704)
- Setup sponsor related info (#730)
- new _run() (#731)
- PPOPolicy training: ERROR: DomainError with NaN: Normal: the condition σ >= zero(σ) is not satisfied. (#739)
- Code Readability (#740)
- MultiThreadEnv not available in ReinforcementLearningZoo (#741)
- ReinforcementLearningExperiment dependencies fail to precompile (#744)
- tanh normalization destabilizes learning with GaussianNetwork (#745)
- Custom Environment Passes RLBase.test_runnable!(env) but infinite hangs and crashes when run. (#757)
- Collect both number of steps and rewards in a single hook (#763)
- Every single environment / experiment crashes with following error: (#766)
- Neural Network Approximator based policies not working (#770)
- "params not defined," "JuliaRL_BasicDQN_CartPole" (#778)
Merged pull requests:
- WIP: Add MPO in zoo (#604) (@HenriDeh)
- Episode reset condition (#621) (@HenriDeh)
- Add a categorical Network (#625) (@HenriDeh)
- Use Trajectories.jl instead (#632) (@findmyway)
- added basic doc for
TDLearner
(#649) (@baedan) - Add
JuliaRL_DQN_CartPole
(#650) (@findmyway) - enable OpenSpiel (#691) (@findmyway)
- Small improvements for TicTacToeEnv (#692) (@jonathan-laurent)
- Update the "how to implement a new algorithm" (#695) (@HenriDeh)
- Fix typo in algorithm implementation docs (#697) (@mplemay)
- add PrioritizedDQN (#698) (@findmyway)
- add QRDQN (#699) (@findmyway)
- add REMDQN (#708) (@findmyway)
- add IQN (#710) (@findmyway)
- checkin Mainifest.toml (#711) (@findmyway)
- CompatHelper: bump compat for "ReinforcementLearningCore" to "0.8" (#712) (@github-actions[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.6" (#713) (@github-actions[bot])
- CompatHelper: bump compat for "ReinforcementLearningZoo" to "0.5" (#714) (@github-actions[bot])
- CompatHelper: bump compat for "AbstractTrees" to "0.4" for package ReinforcementLearningBase (#715) (@github-actions[bot])
- CompatHelper: bump compat for "Functors" to "0.3" for package ReinforcementLearningCore (#717) (@github-actions[bot])
- CompatHelper: bump compat for "UnicodePlots" to "3" for package ReinforcementLearningCore (#718) (@github-actions[bot])
- CompatHelper: bump compat for "ReinforcementLearningCore" to "0.8" for package ReinforcementLearningZoo (#720) (@github-actions[bot])
- CompatHelper: bump compat for "Functors" to "0.3" for package ReinforcementLearningZoo (#721) (@github-actions[bot])
- CompatHelper: add new compat entry for "StableRNGs" at version "1" for package ReinforcementLearningExperiments (#722) (@github-actions[bot])
- CompatHelper: bump compat for "ReinforcementLearning" to "0.10" for package ReinforcementLearningExperiments (#723) (@github-actions[bot])
- add rainbow (#724) (@findmyway)
- Adapted SAC to support MultiThreadedEnv (#726) (@BigFood2307)
- Add the number of episodes (#727) (@ll7)
- docs: add ll7 as a contributor for doc (#728) (@allcontributors[bot])
- Add struct view (#732) (@findmyway)
- add VPG (#733) (@findmyway)
- CompatHelper: add new compat entry for "Distributions" at version "0.25" for package ReinforcementLearningZoo (#734) (@github-actions[bot])
- CompatHelper: add new compat entry for "Distributions" at version "0.25" for package ReinforcementLearningExperiments (#735) (@github-actions[bot])
- fixed hyperlink in readme (#742) (@mplemay)
- docs: add mplemay as a contributor for doc (#743) (@allcontributors[bot])
- Create FUNDING.yml (#746) (@findmyway)
- TRPO (#747) (@baedan)
- CompatHelper: bump compat for "CommonRLSpaces" to "0.2" for package ReinforcementLearningBase (#748) (@github-actions[bot])
- Fix parameter names for AsyncTrajectoryStyle (#749) (@ludvigk)
- Update DoEveryNEpisode hook to new api (#750) (@ludvigk)
- docs: add ludvigk as a contributor for code (#751) (@allcontributors[bot])
- Update TwinNetwork (#752) (@ludvigk)
- Typo in hooks docs (#754) (@kir0ul)
- CommonRLSpace -> DomainSets (#756) (@findmyway)
- Fix typo (#767) (@jeremiahpslewis)
- Fix typo (#768) (@jeremiahpslewis)
- Fix TD Learner so that it handles MultiAgent/Simultaneous with NoOp (#769) (@jeremiahpslewis)
- Bump RLBase compat to 0.11 (#771) (@HenriDeh)
- Remove manifest from the repo (#773) (@HenriDeh)
- import params and gradient (#774) (@HenriDeh)
- fix compat (#775) (@HenriDeh)
- Trying to reimplement experiments (#776) (@HenriDeh)
- Add a developer mode (#777) (@HenriDeh)
- added pettingzoo and one single agent example (#782) (@Mytolo)
- Update mpo.jl (#783) (@HenriDeh)
- Reduce unnecessary array allocations (#785) (@jeremiahpslewis)
- Temporarily disable failing experiment so project tests pass (#787) (@jeremiahpslewis)
- Fix spellcheck errors (#788) (@jeremiahpslewis)
- Bug fixes and dependency bump (#789) (@jeremiahpslewis)
- Pin ReinforcementLearning.jl to pre-refactor versions (#793) (@jeremiahpslewis)
v0.10.1
ReinforcementLearning v0.10.1
Closed issues:
- Support compression? (#102)
- State monitoring and fault tolerance (#101)
- Add card game environments (#196)
- Add Discrete Batch-Constrained Deep Q-learning (#226)
- Add a dedicated multi-dimensional space type (#268)
- PyCall.getindex in module ReinforcementLearningEnvironments conflict warning (#527)
device
method definition overwritten (#530)- StackFrames bug? (#551)
- Small performance improvement (#558)
- Infinite-recursion bug in function
is_discrete_space
when an object of type ClosedInterval is passed (#566) - action_space not defined in tutorial (#569)
- Warning while precompiling RLCore due to kwargs (#575)
- Strange Bug with examples CartPoleEnv and RLBase.test_runnable!(RandomWalk1D) (#579)
- Difficulty Creating a Custom Environment (#581)
- Missing docs for how to implement a new algorithm (#582)
- Donation (#595)
MultiThreadEnv
with custom (continuous) action spaces fails (#596)- PPOCartPole fails, source of error included (#605)
- Bug: Issue with TD3 for multi-dimensional action spaces (#624)
ActionTransformedEnv
doesn't transformlegal_action_space_mask
(#642)- Bug: Previous example from RLZoo now has a bug (#643)
Merged pull requests:
- Fix rlexps (#525) (@findmyway)
- Bump rlenvs (#526) (@findmyway)
- close #527 (#528) (@bhatiaabhinav)
- docs: add bhatiaabhinav as a contributor for bug, code (#529) (@allcontributors[bot])
- Refine the doc and make minor changes of TabularApproximator (#532) (@ShuhuaGao)
- Fix bug in MaskedPPOTrajectory (#533) (@findmyway)
- bugfix with ZeroTo (#534) (@findmyway)
- Revert unexpected change in PPO (#535) (@findmyway)
- Fix 530 (#536) (@findmyway)
- Improves plotting for classical control experiments (#537) (@harwiltz)
- Fix rldatasets (#538) (@findmyway)
- docs: add harwiltz as a contributor for code, doc (#539) (@allcontributors[bot])
- Bump version (#540) (@findmyway)
- fix RLIntro#63 (#541) (@findmyway)
- fix RLIntro#64 (#542) (@findmyway)
- Added a continuous option for CartPoleEnv (#543) (@dylan-asmar)
- docs: add dylan-asmar as a contributor for code (#544) (@allcontributors[bot])
- Bump version (#545) (@findmyway)
- Fix bug in cart pole float32 (#547) (@findmyway)
- Update links to RLIntro (#548) (@findmyway)
- Make experiments GPU compatible (#549) (@findmyway)
- Add compat (#550) (@findmyway)
- Bugfix with cart pole env (#552) (@findmyway)
- make bc gpu compatable (#553) (@findmyway)
- docs: add andreyzhitnikov as a contributor for bug (#554) (@allcontributors[bot])
- Small typo (#555) (@kir0ul)
- docs: add kir0ul as a contributor for doc (#556) (@allcontributors[bot])
- Fix/rand dummy action (#559) (@mo8it)
- Fix warning about kwargs.data (#560) (@mo8it)
- docs: add Mo8it as a contributor for code (#561) (@allcontributors[bot])
- Fix dummy action for continuous action spaces (#562) (@mo8it)
- Fix/rand interval (#563) (@mo8it)
- Remove unneeded method (#564) (@mo8it)
- Fix typo in ospp_final_term_report_210370741/index.md (#565) (@eltociear)
- Fix 566 (#567) (@findmyway)
- Fix documentation for environments (#570) (@blegat)
- docs: add blegat as a contributor for doc (#571) (@allcontributors[bot])
- fix #569 (#573) (@findmyway)
- CompatHelper: bump compat for "ArrayInterface" to "4" for package ReinforcementLearningCore (#574) (@github-actions[bot])
- bump version of RLCore and RLZoo (#576) (@findmyway)
- Update EpsilonGreedyExplorer example (#577) (@kir0ul)
- CompatHelper: bump compat for "FillArrays" to "0.13" for package ReinforcementLearningCore (#583) (@github-actions[bot])
- Default qnetwork initializer (#586) (@HenriDeh)
- docs: add HenriDeh as a contributor for code, doc (#587) (@allcontributors[bot])
- using
act_limit
parameter intarget_actor
(#588) (@NPLawrence) - docs: add NPLawrence as a contributor for code (#589) (@allcontributors[bot])
- CompatHelper: bump compat for "ArrayInterface" to "5" for package ReinforcementLearningCore (#590) (@github-actions[bot])
- Fix documentation typo (#591) (@kir0ul)
- Fixing and generalizing GaussianNetwork (#592) (@HenriDeh)
- Fix typos in docs (#593) (@bileamScheuvens)
- docs: add bileamScheuvens as a contributor for doc (#594) (@allcontributors[bot])
- Add CovGaussianNetwork to work with covariance (#597) (@HenriDeh)
- Fixing Gaussian Network gradient (#598) (@HenriDeh)
- Update Supporting (#599) (@findmyway)
- docs: add harwiltz as a contributor for bug (#601) (@allcontributors[bot])
- Rewrite initialization of
StackFrames
(#602) (@findmyway) - fix test logdetLorU with Float64 (#603) (@HenriDeh)
- fix #605 (#606) (@findmyway)
- docs: add jarbus as a contributor for bug (#607) (@allcontributors[bot])
- Add a reward normalizer (#609) (@HenriDeh)
- cspell add Optimise (#622) (@HenriDeh)
- write doc (#627) (@HenriDeh)
- fix #624 (#628) (@findmyway)
- docs: add tyleringebrand as a contributor for bug (#629) (@allcontributors[bot])
- Update How_to_implement_a_new_algorithm.md (#630) (@HenriDeh)
- add a new notebook (#631) (@findmyway)
- created fallback implementation for legal_action_space_mask (#644) (@baedan)
- update node version (#645) (@findmyway)
- docs: add baedan as a contributor for code (#646) (@allcontributors[bot])
- Tag the latest code as v0.10.1 (#647) (@findmyway)
v0.10.0
ReinforcementLearning v0.10.0
Closed issues:
- In DDPG: Add support for vector actions (#138)
- Add experiments based on offline RL data (#141)
- Train policy with GymEnv (#175)
- SARTTrajectory for SAC (#182)
- PPO related algorithms are broken (#194)
- ERROR: type RandomPolicy has no field policy (#208)
- "Getting Started" too long imo (#210)
- Documentation of environment;
actions
seems not work. (#222) - Documentation of "How to use Tensorboard?":
with_logger
not defined (#223) - Getting figure object; How to get an animation using GR.plot in CartPolEnv (#246)
- The components of Rainbow (#229)
- code in get_started seems to be broken (#233)
- Document how to save/load parameters (#238)
- Workflow of saving trajectory data (#239)
- [Call for Contributors] Summer 2021 of Open Source Promotion Plan (#242)
- Next Release Plan (v0.9) (#247)
- Add ReinforcementLearningDatasets (#253)
- Lack of reproducibility of QRDQN CartPole Experiment. (#281)
- StopAfterNoImprovement hook test fails occasionally (#297)
- Get error when using ReinforcementLearning (#298)
- Problems with PGFPlotsX during the install (#301)
- Plotting CartPole environment in Jupyter (#306)
- Local development environment setup tips causing error (#320)
- Question about PER (#328)
- Docs error in code output (#332)
- Setup a CI for typo (#336)
- double code & dysfunctional master branch when downloading package (#341)
- Precompilation error;
using Plots
makes a conflict (#349) - Problem with running initial tutorial. Using
TabularPolicy()
generates anUndefinedKeyword
error forn_action
(#354) - Question: Clarification on the RL plots generated by the
run()
function (#357) - prob question for QBasedPolicy (#360)
- Can
evaluate
function be used as a component of RLcore? (#369) - problem about precompiling the forked package (#377)
- Question: Can we use packages like
DifferentialEquations.jl
to evolve or model the environment inReinforcementLearning.jl
(#378) - MultiAgentManager does not select correct action space for RockPaperScissorsEnv (#393)
- Add ReinforcementLearningDatasets.jl (#397)
- error: dimension mismatch "cannot broadcast array to have fewer dimensions" (#400)
- SAC policy problems? (#410)
- Add pre-training hook (#411)
- Dead links in documentation (#418)
- Links of
show nbview
badges in RLExperiments are incorrect (#421) - Problem accessing public google cloud storage bucket for RLDatasets.jl (#424)
- Function to access base env through multiple wrapper layers (#425)
- The problem of using GaussianNetwork in gpu (#455)
- Next Release Plan (v0.10) (#460)
- Error in experiment "JuliaRL_DDPG_Pendulum" (#471)
- In Windows, ReinforcementLearningDataset.jl encounter a bug (#485)
- Conditional Testing (#493)
- Inconsistency of the EpsilonGreedyExplorer selection function (#520)
Merged pull requests:
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@LaarsOman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment (#477) (@ShuhuaGao) - docs: add ShuhuaGao as a contributor for code, question (#478) (@allcontributors[bot])
- Chancestyle doc update (#479) (@ShuhuaGao)
- FisherBRC algorithm and update docs (#480) (@pilgrimygy)
- Add the experiment of
MADDPG
. (#481) (@peterchen96) - Add bsuite datasets (#482) (@Mobius1D)
- update report and add reward bonus (#483) (@pilgrimygy)
- Update manifest (#484) (@Mobius1D)
- add GPU support for GaussianNetwork, fix #455 (#486) (@burmecia)
- Update experiments of maddpg (#487) (@peterchen96)
- �Update prob function of
QBasedPolicy.
(#488) (@peterchen96) - Update report. (#489) (@peterchen96)
- Fix gsutil for windows and fix docs (#491) (@Mobius1D)
- add vmpo algorithm (#492) (@burmecia)
- update vae (#494) (@pilgrimygy)
- Add dm datasets (#495) (@Mobius1D)
- Play OpenSpiel envs with NFSP and try to add ED algorithm. (#496) (@peterchen96)
- fix bug (#497) (@pilgrimygy)
- Update BEAR algorithm (#498) (@pilgrimygy)
- More efficient float32 randn (#499) (@albheim)
- Add support for deep ope in RLDatasets.jl (#500) (@Mobius1D)
- update BCQ (#501) (@pilgrimygy)
- update discrete BCQ (#502) (@pilgrimygy)
- update offline RL experiment (#507) (@pilgrimygy)
- Update ED algorithm and the report. (#508) (@peterchen96)
- cpu and gpu test (#509) (@pilgrimygy...
v0.9.0
ReinforcementLearning v0.9.0
Closed issues:
- Roadmap of v0.9 (#65)
- Classic environments in separate package? (#123)
- Add dueling DQN (#137)
- How should ReinforcementLearning.jl be cited ? (#80)
- Alternative handling of max steps in environment (#140)
- Add Highway env (#120)
- Add experiments with GymEnv (#147)
- Replace Travis with github actions (#151)
- AbstractStage docstring doesn't render correctly in docs. (#160)
- List of contributors (#161)
- Return experiment instead of hook only (#173)
- Training mode and testing mode (#179)
- AbstractEnv (#181)
- define environment of FULL_ACTION_SET (#184)
- CircularArraySLARTTrajectory instance is not of type CircularArraySLARTTrajectory (#185)
- Is hook the same thing as "callback"? (#190)
- Use @threads instead of @sync + @Spawn in MultiThreadEnv? (#191)
- Blog custom env link typo (#192)
- Separate envs from algos in Zoo? (#197)
- Why "examples"? (#198)
- WandB integration? (#201)
- Add default implementations for AbstractEnvWrapper (#202)
- Add configuration in DQNLearner to enable double-dqn by default (#205)
- Why split repos? (#209)
- PreActStage clarification (#212)
- What's a "trace"? (#213)
- Continuous time supported? (#215)
- Docs looks ugly in dark mode (#217)
- Julia 1.6.0 dependency problem with
ReinforcementLearningBase
/RLBase (#221) - Docstring of DoEveryNStep (#225)
- Update dependency to [email protected] and resolve type piracy of findmax (#227)
- IQN is broken with [email protected] (#228)
- Source links in documentation directs to julia repo (#231)
- PPO strange behaviour from having actions as one element arrays instead of scalar (#234)
- SAC and GaussianNetwork (#236)
- Precompilation prohibitively long (#237)
- An explanation of "how to train policy (agent)" such as
Basic_DQN
would be valuable (#240) - How to guarantee the environment's reproducibility? (#241)
- Cannot use RLBase.action_space etc. when writing my own environment (#244)
- ReinforcementLearningZoo.jl experiments (#245)
- How about making this package compatible with DifferentialEquations.jl? (#249)
- PPO and multi dimensional actions spaces (#251)
- Incompatibility with CSVFiles.jl (#256)
- [RLEnvs] easy access of the length of an action vector (dimension of action space) (#257)
- Cannot add
LinearAlgebra
(#258) - What's the
checkpoints
? (#261) - PPO is broken when using CUDA (#280)
- Reinforcement Learning.jl in a RTS (#291)
Merged pull requests:
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
v0.8.0
ReinforcementLearning v0.8.0
Closed issues:
- Document basic environments (#129)
- Improve interfaces for model exploration and hyperparameter optimization (#28)
- Support SEED RL (SCALABLE AND EFFICIENT DEEP-RL ) (#62)
- Rename
AbstractAgent
toAbstractPolicy
(#111) - Add a stop condition to terminate the experiment after reaching reward threashold (#112)
- ACME RL lib by deepmind (#85)
- Definition of a policy (#86)
- Add remote trajectories (#87)
- Base.convert method for DiscreteSpace (#104)
- Action Space Meaning (#88)
- Base.in method for EmptySpace (#105)
- Renaming get_terminal to isterminated (#106)
- Requesting more informative field names for SharedTrajectory (#113)
- Suggestion: More informative name for FullActionSet & MinimalActionSet (#107)
- Returning an
AbstractSpace
object usingget_actions
(#108) - Split experiments into separate files (#145)
- Add project.toml for tests (#146)
- Docs build error (#91)
- Split out Trajectory & CircularArrayBuffer as independent packages (#114)
- Requesting explanation for better performance at ... (#115)
- Add an extra mode when evaluating agent (#116)
- Why are wrapper environments a part of RLBase instead of RLCore (say)? (#109)
- The names of keyword arguments in Trajectory is kind of misunderstanding (#117)
- Check compatibility between agent and environments (#118)
- Behaviour for hooks for RewardOverridenEnv (#119)
- StopAfterEpisode with custom DQNL errors beyond a particular Episode Count (#96)
ERROR: UndefVarError: NNlib not defined
while loading agent (#110)- Use JLSO for (de)serialization? (#97)
- Setup github actions (#98)
- Fails to load trajectory (#150)
- Test error in ReinforcementLearningEnvironments.jl (#152)
- Move preallocations in MultiThreadEnv from global to local (#153)
- remove @views (#155)
- error in save & load ElasticCompactSARTSATrajectory (#156)
- add early stopping in src\core\stop_conditions.jl (#157)
- add time stamp in load & save function, in file src\components\agents\agent.jl (#158)
- policies in GPU can not be saved || loaded (#159)
- code formatting (#165)
- Purpose of CommonRLInterface (#166)
- Moving example environments from RLBase to RLEnvs? (#167)
- Keeping prefix
get_
in method names likeget_reward
(#168) - Currently getting an ambiguous method error in ReinforcementLearningCore v0.5.1 (#171)
- TD3 Implementation (#174)
- Travis CI Credits (#178)
- Unrecognized symbols (#180)
Merged pull requests:
- update dependency (#177) (@findmyway)
v0.7.0
ReinforcementLearning v0.7.0
Closed issues:
v0.6.0
ReinforcementLearning v0.6.0
Make environments transparent to agents/policies.
Mooncake release may be delayed.
Closed issues:
Merged pull requests:
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
v0.5.0
ReinforcementLearning v0.5.0
Closed issues:
- Box2D environment (#2)
- bullet3 environment (#7)
- loadenvironment error (#19)
- Support alternative deep learning libraries (#20)
- Random Thoughts on v0.3.0 (#24)
- Prioritized DQN (#27)
- A2C (#32)
- Add built-in support for TensorBoard (#35)
- Add checkpoints (#36)
- Improve code coverage (#40)
- AbstractActionSelector not exported (#41)
- Params empty - no tracking (#43)
- Add reproducible examples for Atari environments (#44)
- StopAfterEpisode with progress meter (#51)
- Support julia 1.4 (#63)
Merged pull requests:
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
Preview of v0.4.0
v0.4.0-beta Install TagBot as a GitHub Action (#53)
Preview of v0.4.0
This is a preview of v0.4.0
.
What we have for now:
- Tabular methods are well tested.
- Some simple value based methods are implemented.
What are missing:
- GPU support
- Some other DRL methods proposed in the issues with tag
v0.4.0