v0.10.1
github-actions
released this
04 Jun 16:10
·
141 commits
to master
since this release
ReinforcementLearning v0.10.1
Closed issues:
- Support compression? (#102)
- State monitoring and fault tolerance (#101)
- Add card game environments (#196)
- Add Discrete Batch-Constrained Deep Q-learning (#226)
- Add a dedicated multi-dimensional space type (#268)
- PyCall.getindex in module ReinforcementLearningEnvironments conflict warning (#527)
device
method definition overwritten (#530)- StackFrames bug? (#551)
- Small performance improvement (#558)
- Infinite-recursion bug in function
is_discrete_space
when an object of type ClosedInterval is passed (#566) - action_space not defined in tutorial (#569)
- Warning while precompiling RLCore due to kwargs (#575)
- Strange Bug with examples CartPoleEnv and RLBase.test_runnable!(RandomWalk1D) (#579)
- Difficulty Creating a Custom Environment (#581)
- Missing docs for how to implement a new algorithm (#582)
- Donation (#595)
MultiThreadEnv
with custom (continuous) action spaces fails (#596)- PPOCartPole fails, source of error included (#605)
- Bug: Issue with TD3 for multi-dimensional action spaces (#624)
ActionTransformedEnv
doesn't transformlegal_action_space_mask
(#642)- Bug: Previous example from RLZoo now has a bug (#643)
Merged pull requests:
- Fix rlexps (#525) (@findmyway)
- Bump rlenvs (#526) (@findmyway)
- close #527 (#528) (@bhatiaabhinav)
- docs: add bhatiaabhinav as a contributor for bug, code (#529) (@allcontributors[bot])
- Refine the doc and make minor changes of TabularApproximator (#532) (@ShuhuaGao)
- Fix bug in MaskedPPOTrajectory (#533) (@findmyway)
- bugfix with ZeroTo (#534) (@findmyway)
- Revert unexpected change in PPO (#535) (@findmyway)
- Fix 530 (#536) (@findmyway)
- Improves plotting for classical control experiments (#537) (@harwiltz)
- Fix rldatasets (#538) (@findmyway)
- docs: add harwiltz as a contributor for code, doc (#539) (@allcontributors[bot])
- Bump version (#540) (@findmyway)
- fix RLIntro#63 (#541) (@findmyway)
- fix RLIntro#64 (#542) (@findmyway)
- Added a continuous option for CartPoleEnv (#543) (@dylan-asmar)
- docs: add dylan-asmar as a contributor for code (#544) (@allcontributors[bot])
- Bump version (#545) (@findmyway)
- Fix bug in cart pole float32 (#547) (@findmyway)
- Update links to RLIntro (#548) (@findmyway)
- Make experiments GPU compatible (#549) (@findmyway)
- Add compat (#550) (@findmyway)
- Bugfix with cart pole env (#552) (@findmyway)
- make bc gpu compatable (#553) (@findmyway)
- docs: add andreyzhitnikov as a contributor for bug (#554) (@allcontributors[bot])
- Small typo (#555) (@kir0ul)
- docs: add kir0ul as a contributor for doc (#556) (@allcontributors[bot])
- Fix/rand dummy action (#559) (@mo8it)
- Fix warning about kwargs.data (#560) (@mo8it)
- docs: add Mo8it as a contributor for code (#561) (@allcontributors[bot])
- Fix dummy action for continuous action spaces (#562) (@mo8it)
- Fix/rand interval (#563) (@mo8it)
- Remove unneeded method (#564) (@mo8it)
- Fix typo in ospp_final_term_report_210370741/index.md (#565) (@eltociear)
- Fix 566 (#567) (@findmyway)
- Fix documentation for environments (#570) (@blegat)
- docs: add blegat as a contributor for doc (#571) (@allcontributors[bot])
- fix #569 (#573) (@findmyway)
- CompatHelper: bump compat for "ArrayInterface" to "4" for package ReinforcementLearningCore (#574) (@github-actions[bot])
- bump version of RLCore and RLZoo (#576) (@findmyway)
- Update EpsilonGreedyExplorer example (#577) (@kir0ul)
- CompatHelper: bump compat for "FillArrays" to "0.13" for package ReinforcementLearningCore (#583) (@github-actions[bot])
- Default qnetwork initializer (#586) (@HenriDeh)
- docs: add HenriDeh as a contributor for code, doc (#587) (@allcontributors[bot])
- using
act_limit
parameter intarget_actor
(#588) (@NPLawrence) - docs: add NPLawrence as a contributor for code (#589) (@allcontributors[bot])
- CompatHelper: bump compat for "ArrayInterface" to "5" for package ReinforcementLearningCore (#590) (@github-actions[bot])
- Fix documentation typo (#591) (@kir0ul)
- Fixing and generalizing GaussianNetwork (#592) (@HenriDeh)
- Fix typos in docs (#593) (@bileamScheuvens)
- docs: add bileamScheuvens as a contributor for doc (#594) (@allcontributors[bot])
- Add CovGaussianNetwork to work with covariance (#597) (@HenriDeh)
- Fixing Gaussian Network gradient (#598) (@HenriDeh)
- Update Supporting (#599) (@findmyway)
- docs: add harwiltz as a contributor for bug (#601) (@allcontributors[bot])
- Rewrite initialization of
StackFrames
(#602) (@findmyway) - fix test logdetLorU with Float64 (#603) (@HenriDeh)
- fix #605 (#606) (@findmyway)
- docs: add jarbus as a contributor for bug (#607) (@allcontributors[bot])
- Add a reward normalizer (#609) (@HenriDeh)
- cspell add Optimise (#622) (@HenriDeh)
- write doc (#627) (@HenriDeh)
- fix #624 (#628) (@findmyway)
- docs: add tyleringebrand as a contributor for bug (#629) (@allcontributors[bot])
- Update How_to_implement_a_new_algorithm.md (#630) (@HenriDeh)
- add a new notebook (#631) (@findmyway)
- created fallback implementation for legal_action_space_mask (#644) (@baedan)
- update node version (#645) (@findmyway)
- docs: add baedan as a contributor for code (#646) (@allcontributors[bot])
- Tag the latest code as v0.10.1 (#647) (@findmyway)