v0.10.0
github-actions
released this
08 Oct 12:20
·
223 commits
to master
since this release
ReinforcementLearning v0.10.0
Closed issues:
- In DDPG: Add support for vector actions (#138)
- Add experiments based on offline RL data (#141)
- Train policy with GymEnv (#175)
- SARTTrajectory for SAC (#182)
- PPO related algorithms are broken (#194)
- ERROR: type RandomPolicy has no field policy (#208)
- "Getting Started" too long imo (#210)
- Documentation of environment;
actions
seems not work. (#222) - Documentation of "How to use Tensorboard?":
with_logger
not defined (#223) - Getting figure object; How to get an animation using GR.plot in CartPolEnv (#246)
- The components of Rainbow (#229)
- code in get_started seems to be broken (#233)
- Document how to save/load parameters (#238)
- Workflow of saving trajectory data (#239)
- [Call for Contributors] Summer 2021 of Open Source Promotion Plan (#242)
- Next Release Plan (v0.9) (#247)
- Add ReinforcementLearningDatasets (#253)
- Lack of reproducibility of QRDQN CartPole Experiment. (#281)
- StopAfterNoImprovement hook test fails occasionally (#297)
- Get error when using ReinforcementLearning (#298)
- Problems with PGFPlotsX during the install (#301)
- Plotting CartPole environment in Jupyter (#306)
- Local development environment setup tips causing error (#320)
- Question about PER (#328)
- Docs error in code output (#332)
- Setup a CI for typo (#336)
- double code & dysfunctional master branch when downloading package (#341)
- Precompilation error;
using Plots
makes a conflict (#349) - Problem with running initial tutorial. Using
TabularPolicy()
generates anUndefinedKeyword
error forn_action
(#354) - Question: Clarification on the RL plots generated by the
run()
function (#357) - prob question for QBasedPolicy (#360)
- Can
evaluate
function be used as a component of RLcore? (#369) - problem about precompiling the forked package (#377)
- Question: Can we use packages like
DifferentialEquations.jl
to evolve or model the environment inReinforcementLearning.jl
(#378) - MultiAgentManager does not select correct action space for RockPaperScissorsEnv (#393)
- Add ReinforcementLearningDatasets.jl (#397)
- error: dimension mismatch "cannot broadcast array to have fewer dimensions" (#400)
- SAC policy problems? (#410)
- Add pre-training hook (#411)
- Dead links in documentation (#418)
- Links of
show nbview
badges in RLExperiments are incorrect (#421) - Problem accessing public google cloud storage bucket for RLDatasets.jl (#424)
- Function to access base env through multiple wrapper layers (#425)
- The problem of using GaussianNetwork in gpu (#455)
- Next Release Plan (v0.10) (#460)
- Error in experiment "JuliaRL_DDPG_Pendulum" (#471)
- In Windows, ReinforcementLearningDataset.jl encounter a bug (#485)
- Conditional Testing (#493)
- Inconsistency of the EpsilonGreedyExplorer selection function (#520)
Merged pull requests:
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@LaarsOman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment (#477) (@ShuhuaGao) - docs: add ShuhuaGao as a contributor for code, question (#478) (@allcontributors[bot])
- Chancestyle doc update (#479) (@ShuhuaGao)
- FisherBRC algorithm and update docs (#480) (@pilgrimygy)
- Add the experiment of
MADDPG
. (#481) (@peterchen96) - Add bsuite datasets (#482) (@Mobius1D)
- update report and add reward bonus (#483) (@pilgrimygy)
- Update manifest (#484) (@Mobius1D)
- add GPU support for GaussianNetwork, fix #455 (#486) (@burmecia)
- Update experiments of maddpg (#487) (@peterchen96)
- �Update prob function of
QBasedPolicy.
(#488) (@peterchen96) - Update report. (#489) (@peterchen96)
- Fix gsutil for windows and fix docs (#491) (@Mobius1D)
- add vmpo algorithm (#492) (@burmecia)
- update vae (#494) (@pilgrimygy)
- Add dm datasets (#495) (@Mobius1D)
- Play OpenSpiel envs with NFSP and try to add ED algorithm. (#496) (@peterchen96)
- fix bug (#497) (@pilgrimygy)
- Update BEAR algorithm (#498) (@pilgrimygy)
- More efficient float32 randn (#499) (@albheim)
- Add support for deep ope in RLDatasets.jl (#500) (@Mobius1D)
- update BCQ (#501) (@pilgrimygy)
- update discrete BCQ (#502) (@pilgrimygy)
- update offline RL experiment (#507) (@pilgrimygy)
- Update ED algorithm and the report. (#508) (@peterchen96)
- cpu and gpu test (#509) (@pilgrimygy)
- Fix dispatch for is_discrete_space (#510) (@johannes-fischer)
- docs: add johannes-fischer as a contributor for code (#511) (@allcontributors[bot])
- update report (#512) (@pilgrimygy)
- update report (#513) (@peterchen96)
- Fix random net init in sac example (#514) (@albheim)
- WIP to implement FQE (#515) (@Mobius1D)
- OSPP Report for RLDatasets.jl (#516) (@Mobius1D)
- Update report (#518) (@Mobius1D)
- Update reward wrappers to be more consistent (#519) (@albheim)
- fixed findmax unconsistency (#521) (@3rdCore)
- docs: add 3rdCore as a contributor for bug, code (#522) (@allcontributors[bot])
- close #493 (#523) (@findmyway)
- Update compat & version (#524) (@findmyway)