Releases: JuliaReinforcementLearning/ReinforcementLearning.jl
Releases · JuliaReinforcementLearning/ReinforcementLearning.jl
ReinforcementLearningCore-v0.15.5
ReinforcementLearningCore ReinforcementLearningCore-v0.15.5
Diff since ReinforcementLearningCore-v0.15.4
Merged pull requests:
- Move UnicodePlots to extension (#1088) (@jeremiahpslewis)
- Bump RLCore to v0.15.5 (#1089) (@jeremiahpslewis)
ReinforcementLearningFarm-v0.0.3
ReinforcementLearningFarm ReinforcementLearningFarm-v0.0.3
Merged pull requests:
- Fix deprecations (#10) (@femtocleaner[bot])
- implement epsilon-greedy policy with parametric type (#12) (@jbrea)
- improve docs (#13) (@jbrea)
- refactor policies (#15) (@jbrea)
- Add ReinforcementLearningBase as dependent (#16) (@jbrea)
- fix examples (#18) (@jbrea)
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- update dependency (#177) (@findmyway)
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@Nthman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment (#477) (@shuh...
ReinforcementLearningCore-v0.15.4
ReinforcementLearningCore ReinforcementLearningCore-v0.15.4
Diff since ReinforcementLearningCore-v0.15.3
Merged pull requests:
- Fix RLEnvs version (#1076) (@jeremiahpslewis)
- Fix TagBot (#1077) (@jeremiahpslewis)
- Fix docs homepage (#1082) (@michalrzak)
- Jpsl/update flux (#1086) (@jeremiahpslewis)
ReinforcementLearningEnvironments-v0.9.1
ReinforcementLearningEnvironments ReinforcementLearningEnvironments-v0.9.1
Diff since ReinforcementLearningEnvironments-v0.9.0
Merged pull requests:
- Add missing Flux compat (#1059) (@jeremiahpslewis)
- Fix docs / website build (#1064) (@jeremiahpslewis)
- Correct Pendulum x-y coordinates (#1065) (@HenriDeh)
- Make QBasedPolicy general for AbstractLearner s (#1069) (@dharux)
- Fix hooks for multiplayer case (#1071) (@jeremiahpslewis)
- Fix doc build errors (#1072) (@jeremiahpslewis)
- bump rlcore version (#1073) (@jeremiahpslewis)
- Make
FluxApproximator
work withQBasedPolicy
(#1075) (@jeremiahpslewis) - Fix RLEnvs version (#1076) (@jeremiahpslewis)
Closed issues:
- Next Release Plan (v0.11) (#614)
- Package Stabilization Plan (#792)
- test/runtests.jl empty (+ arch discussion) (#843)
- policy(env) returns no legal action -inf initialized Q-table (#852)
- Refactor CI into separate Workflows per package (and separate codecov projects per package) (#869)
- Add deprecation warnings to non-refactored policies (#892)
- Vectorized environments (#908)
- Loading a Gym Environment (#912)
- PPO with MaskedPPOTrajectory (#917)
- Devmode is not working (#918)
- TD3 Policy unable to handle environments with multidimensional action spaces (#951)
- Spin off core packages (#960)
- experiments failed (#982)
- Breaking the tutorial by getting TotalRewardPerEpisode out of sync with the stopping condition in a
run
call (#1000) - Transfer Algorithms to RLFarm (#1028)
- Update Buildkite script for gpu testing so it's sub package compatible (#1030)
- Website: A practical introduction to RL: Does not introduce, source code is broken (#1036)
- ElasticArraySARTSTraces does not record the trajectories of
MountainCarEnv()
correctly (#1067) - Algorithm implementations (#1070)
- No method matching iterate ArrayProductDomain (#1074)
ReinforcementLearningCore-v0.15.3
ReinforcementLearningCore ReinforcementLearningCore-v0.15.3
Diff since ReinforcementLearningCore-v0.15.2
Merged pull requests:
- Make
FluxApproximator
work withQBasedPolicy
(#1075) (@jeremiahpslewis)
ReinforcementLearningBase-v0.13.1
ReinforcementLearningBase ReinforcementLearningBase-v0.13.1
Diff since ReinforcementLearningBase-v0.13.0
Merged pull requests:
- Add missing Flux compat (#1059) (@jeremiahpslewis)
- Fix docs / website build (#1064) (@jeremiahpslewis)
- Correct Pendulum x-y coordinates (#1065) (@HenriDeh)
- Make QBasedPolicy general for AbstractLearner s (#1069) (@dharux)
- Fix hooks for multiplayer case (#1071) (@jeremiahpslewis)
- Fix doc build errors (#1072) (@jeremiahpslewis)
- bump rlcore version (#1073) (@jeremiahpslewis)
- Make
FluxApproximator
work withQBasedPolicy
(#1075) (@jeremiahpslewis)
Closed issues:
- Next Release Plan (v0.11) (#614)
- Package Stabilization Plan (#792)
- test/runtests.jl empty (+ arch discussion) (#843)
- policy(env) returns no legal action -inf initialized Q-table (#852)
- Refactor CI into separate Workflows per package (and separate codecov projects per package) (#869)
- Add deprecation warnings to non-refactored policies (#892)
- Vectorized environments (#908)
- Loading a Gym Environment (#912)
- PPO with MaskedPPOTrajectory (#917)
- Devmode is not working (#918)
- TD3 Policy unable to handle environments with multidimensional action spaces (#951)
- Spin off core packages (#960)
- experiments failed (#982)
- Breaking the tutorial by getting TotalRewardPerEpisode out of sync with the stopping condition in a
run
call (#1000) - Transfer Algorithms to RLFarm (#1028)
- Update Buildkite script for gpu testing so it's sub package compatible (#1030)
- Website: A practical introduction to RL: Does not introduce, source code is broken (#1036)
- ElasticArraySARTSTraces does not record the trajectories of
MountainCarEnv()
correctly (#1067) - Algorithm implementations (#1070)
ReinforcementLearningCore-v0.15.2
ReinforcementLearningCore ReinforcementLearningCore-v0.15.2
Diff since ReinforcementLearningCore-v0.15.1
Merged pull requests:
- Make QBasedPolicy general for AbstractLearner s (#1069) (@dharux)
- bump rlcore version (#1073) (@jeremiahpslewis)
ReinforcementLearningCore-v0.15.1
ReinforcementLearningCore ReinforcementLearningCore-v0.15.1
Diff since ReinforcementLearningCore-v0.15.0
Merged pull requests:
- Add missing Flux compat (#1059) (@jeremiahpslewis)
- Fix docs / website build (#1064) (@jeremiahpslewis)
- Correct Pendulum x-y coordinates (#1065) (@HenriDeh)
- Fix hooks for multiplayer case (#1071) (@jeremiahpslewis)
- Fix doc build errors (#1072) (@jeremiahpslewis)
Closed issues:
- Next Release Plan (v0.11) (#614)
- Package Stabilization Plan (#792)
- test/runtests.jl empty (+ arch discussion) (#843)
- policy(env) returns no legal action -inf initialized Q-table (#852)
- Refactor CI into separate Workflows per package (and separate codecov projects per package) (#869)
- Add deprecation warnings to non-refactored policies (#892)
- Vectorized environments (#908)
- Loading a Gym Environment (#912)
- PPO with MaskedPPOTrajectory (#917)
- Devmode is not working (#918)
- TD3 Policy unable to handle environments with multidimensional action spaces (#951)
- Spin off core packages (#960)
- experiments failed (#982)
- Breaking the tutorial by getting TotalRewardPerEpisode out of sync with the stopping condition in a
run
call (#1000) - Transfer Algorithms to RLFarm (#1028)
- Update Buildkite script for gpu testing so it's sub package compatible (#1030)
- Website: A practical introduction to RL: Does not introduce, source code is broken (#1036)
- ElasticArraySARTSTraces does not record the trajectories of
MountainCarEnv()
correctly (#1067) - Algorithm implementations (#1070)
v0.11.0
ReinforcementLearning v0.11.0
Merged pull requests:
- Reactivate some tests for RLExperiments (#790) (@jeremiahpslewis)
- Drop RL.jl as dependency from Experiments (#795) (@jeremiahpslewis)
- Fix compat for RLBase (#796) (@jeremiahpslewis)
- Fix RLCore version, prep for bump (#797) (@jeremiahpslewis)
- Add reexport compat (#798) (@jeremiahpslewis)
- Bump compat helper (#799) (@jeremiahpslewis)
- Fix IntervalSets compat for RLEnvironments (#800) (@jeremiahpslewis)
- Bump RLZoo.jl version for release (#815) (@jeremiahpslewis)
- Fix RLExperiments compat (#816) (@jeremiahpslewis)
- Expand RLZoo compat (#817) (@jeremiahpslewis)
- Bump RLExperiments, require 0.11 (#818) (@jeremiahpslewis)
- Pin ReinforcementLearningZoo.jl to 0.6 in RLExperiments (#819) (@jeremiahpslewis)
- Drop RL.jl from CompatHelper (until refactor complete) (#824) (@jeremiahpslewis)
- Bump Github Actions cache version (#825) (@jeremiahpslewis)
- Basic allocation fixes for RandomWalk / RandomPolicy (#827) (@jeremiahpslewis)
- Bump CI.yml GitHub action versions (#828) (@jeremiahpslewis)
- Add tests, improve performance of RewardsPerEpisode (#829) (@jeremiahpslewis)
- Refactor and add tests to TotalBatchRewardPerEpisode (#830) (@jeremiahpslewis)
- Tests, refactor for TimePerStep (#831) (@jeremiahpslewis)
- DoEveryNStep tests, performance tweaks (#832) (@jeremiahpslewis)
- Add DoOnExit test (#833) (@jeremiahpslewis)
- Expand PR Template (#835) (@jeremiahpslewis)
- Fix branch name (master -> main) (#837) (@jeremiahpslewis)
- Add test_noop! to remaining hooks (#840) (@jeremiahpslewis)
- Make TimePerStep test robust (#841) (@jeremiahpslewis)
- Reactivate docs (#842) (@jeremiahpslewis)
- Add activate_devmode!() explanation to tips.md (#845) (@jeremiahpslewis)
- Bump compat of RL.jl to 0.11.0-dev (#846) (@jeremiahpslewis)
- add kwargs to agent (#847) (@HenriDeh)
- Gaussian network refactor and tests (#849) (@HenriDeh)
- Agent Refactor (#850) (@jeremiahpslewis)
- Bump RLCore (#851) (@jeremiahpslewis)
- Include codecov in CI (#854) (@HenriDeh)
- Fix a typo in MPO (#855) (@HenriDeh)
- DoEvery should not trigger on t = 1 (#856) (@HenriDeh)
- update CI Julia version (#857) (@jeremiahpslewis)
- Tweak CI to check on dep changes (#858) (@HenriDeh)
- CompatHelper: bump compat for FillArrays to 1 for package ReinforcementLearningCore, (keep existing compat) (#859) (@github-actions[bot])
- MultiAgent Proposal (#861) (@jeremiahpslewis)
- CompatHelper: add new compat entry for ReinforcementLearningCore at version 0.9 for package ReinforcementLearningEnvironments, (keep existing compat) (#865) (@github-actions[bot])
- Multiplayer Fixes (Clean up errors) (#867) (@jeremiahpslewis)
- Added a section to the home page about getting help for Reinforcement… (#868) (@LooseTerrifyingSpaceMonkey)
- Bump StatsBase compat (#873) (@jeremiahpslewis)
- ComposedHooks, MultiHook fixes (#874) (@jeremiahpslewis)
- Fix RLEnvs compat (#875) (@jeremiahpslewis)
- Add back ComposedStop (#876) (@jeremiahpslewis)
- Bump RLBase to v0.11.1 (#877) (@jeremiahpslewis)
- Further refinements (#879) (@jeremiahpslewis)
- Use multiple dispatch / methods plan! and act! (#880) (@jeremiahpslewis)
- RLCore.update! -> Base.push! API change (#884) (@jeremiahpslewis)
- Add compat for CommonRLInterface (#886) (@jeremiahpslewis)
- Fix hook issues (#887) (@jeremiahpslewis)
- CompatHelper: bump compat for ReinforcementLearningZoo to 0.6 for package ReinforcementLearningExperiments, (keep existing compat) (#888) (@github-actions[bot])
- Stacknamespace (#889) (@HenriDeh)
- allow more recent versions (#890) (@HenriDeh)
- Fix stack (#891) (@HenriDeh)
- CompatHelper: add new compat entry for DelimitedFiles at version 1 for package ReinforcementLearningEnvironments, (keep existing compat) (#894) (@github-actions[bot])
- Update implement new alg docs (#896) (@jeremiahpslewis)
- NFQ (#897) (@CasBex)
- fixed problem with sequential multi agent envs (#898) (@Mytolo)
- Sketch out optimise! refactor (#899) (@jeremiahpslewis)
- Bug fix optimise! (#902) (@jeremiahpslewis)
- Breaking changes to optimise! interface: Bump RLCore to v0.11 and RLZoo to v0.8 (#903) (@jeremiahpslewis)
- Swap out rng code (#905) (@jeremiahpslewis)
- CompatHelper: bump compat for NNlib to 0.9 for package ReinforcementLearningZoo, (keep existing compat) (#906) (@github-actions[bot])
- Fix dispatch and update documentation (#907) (@HenriDeh)
- QBasedPolicy optimise! forwards to learner. (#909) (@HenriDeh)
- Bump RLZoo version for NNlib (#911) (@jeremiahpslewis)
- Add performance testing run loop (#914) (@jeremiahpslewis)
- Fix Timer bug (#915) (@jeremiahpslewis)
- couple of improvements to MPO (#919) (@HenriDeh)
- Rework the run loop (#921) (@HenriDeh)
- adjusted pettingzoo to PettingZooEnv simultaneous environment more co… (#925) (@Mytolo)
- fixed devmode / project files (#932) (@Mytolo)
- fixed DQNLearner Gpu isse (#933) (@Mytolo)
- fixing prob. /w symbol/ string correspondence (#934) (@Mytolo)
- Bump flux compat (#935) (@jeremiahpslewis)
- Reduce find_all_max allocations and increase speed based on chatgpt s… (#938) (@jeremiahpslewis)
- Add Buildkite / GPU tests (#942) (@jeremiahpslewis)
- Add RLZoo and RLExperiments to Buildkite (#943) (@jeremiahpslewis)
- Drop deprecated @provide interface (#944) (@jeremiahpslewis)
- CI Improvements (#946) (@jeremiahpslewis)
- Github Actions Fixes (#947) (@jeremiahpslewis)
- gpu updates RLExperiments, RLZoo (#949) (@jeremiahpslewis)
- Bump RLCore version (#950) (@jeremiahpslewis)
- Refactor TRPO and VPG with EpisodesSampler (#952) (@HenriDeh)
- Fix TotalRewardPerEpisode bug (#953) (@jeremiahpslewis)
- update docs to loop refactor (#955) (@HenriDeh)
- CompatHelper: bump compat for ReinforcementLearningCore to 0.13 for package ReinforcementLearningEnvironments, (keep existing compat) (#956) (@github-actions[bot])
- CompatHelper: bump compat for ReinforcementLearningCore to 0.13 for package ReinforcementLearningZoo, (keep existing compat) (#957) (@github-actions[bot])
- CompatHelper: bump compat for ReinforcementLearningCore to 0.13 for package ReinforcementLearningExperiments, (keep existing compat) (#958) (@github-actions[bot])
- CompatHelper: add new compat entry for cuDNN at version 1 for package ReinforcementLearningCore, (keep existing compat) (#962) (@github-actions[bot])
- CompatHelper: add new compat entry for cuDNN at version 1 for package ReinforcementLearningZoo, (keep existing compat) (#963) (@github-actions[bot])
- CompatHelper: add new compat entry for cuDNN at version 1 for package ReinforcementLearningExperiments, (keep existing compat) (#964) (@github-actions[bot])
- TargetNetwork (#966) (@HenriDeh)
- CompatHelper: bump compat for GPUArrays to 9 for package ReinforcementLearningCore, (keep existing compat) (#969) (@github-actions[bot])
- Prioritised DQN GPU (#974) (@CasBex)
- CompatHelper: bump compat for ReinforcementLearningCore to 0.13 for package ReinforcementLearningZoo, (keep existing compat) (#975) (@github-actions[bot])
- CompatHelper: bump compat for ReinforcementLearningZoo to 0.8 for package ReinforcementLearningExperiments, (keep existing compat) (#976) (@github-actions[bot])
- CompatHelper: bump compat for ReinforcementLearningCore to 0.13 for package ReinforcementLearningExperiments, (keep existing compat) (#977) (@github-actions[bot])
- Nfq refactor (#980) (@CasBex)
- Fix and refactor SAC (#985) (@HenriDeh)
- CompatHelper: bump compat for DomainSets to 0.7 for package ReinforcementLearningBase, (keep existing compat) (#986) (@github-actions[bot])
- remove rlenv dep for tests (#989) (@HenriDeh)
- CompatHelper: add new compat entry for CUDA at version 5 for package ReinforcementLearningExperiments, (keep existing compat) (#991) (@github-actions[bot])
- CompatHelper: add new compat entry for IntervalSets at version 0.7 for package ReinforcementLearningExperiments, (keep existing compat) (#994) (@github-actions[bot])
- Conservative Q-Learning (#995) (@HenriDeh)
- CompatHelper: add new compat entry for Parsers at version 2 for package ReinforcementLearningCore, (keep existing compat) (#997) (@github-actions[bot])
- CompatHelper: add new compat entry for MLUtils at version 0.4 for package ReinforcementLearningZoo, (keep existing compat) (#998) (@github-actions[bot])
- CompatHelper: add new compat entry for Statistics at version 1 for package ReinforcementLearningCore, (keep existing compat) (#999) (@github-actions[bot])
- Update CQL_SAC.jl (#1003) (@HenriDeh)
- Bump tj-actions/changed-files from 35 to 41 in /.github/workflows (#1006) (@dependabot[bot])
- Make it compatible with Adapt 4 and Metal 1 (#1008) (@joelreymont)
- Bump RLCore, RLEnv (#1012) (@jeremiahpslewis)
- Fix PPO per #1007 (#1013) (@jeremiahpslewis)
- Fix RLCore version (#1018) (@jeremiahpslewis)
- Add Devcontainer, handle DomainSets 0.7 (#1019) (@jeremiahpslewis)
- Initial GPUArray transition (#1020) (@jeremiahpslewis)
- Update TagBot.yml for subprojects (#1021) (@jeremiahpslewis)
- Fix offline agent test (#1025) (@joelreymont)
- Fix spell check CI errors (#1027) (@joelreymont)
- GPU Code Migration Part 2.1 (#1029) (@jeremiahpslewis)
- Bump RLZoo to v0.8 (#1031) (@jeremiahpslewis)
- Fix RLZoo version (#1032) (@jeremiahpslewis)
- Drop devmode, prepare RL.jl v0.11 for release (#1035) (@jeremiahpslewis)
- Update docs script for new 'limited' RL.jl release (#1038) (@jeremiahpslewis)
- Tabular Approximator fixes (pre v0.11 changes) (#1040) (@jeremiahpslewis)
- Swap RLZoo for RLFarm in CI, drop RLExperiments (#1041) (@jeremiahpslewis)
- Buildkite tweaks for monorepo (#1042) (@jeremiahpslewis)
- Drop archived projects (#1043) (@jeremiahpslewis)
- Simplify Experiment code after dropping RLExperiment (#1044) (@jeremiahpslewis)
- Fix code coverage scope so it ignores test dir (#1045) (@jeremiahpslewis)
- Fix reset and stop conditions (#1046) (@jeremiahpslewis)
- Drop Functors and use Flux.@layer (#1048) (@je...
ReinforcementLearningEnvironments-v0.9.0
ReinforcementLearningEnvironments ReinforcementLearningEnvironments-v0.9.0
Diff since ReinforcementLearningEnvironments-v0.8.8
Merged pull requests:
- Bump RLZoo to v0.8 (#1031) (@jeremiahpslewis)
- Fix RLZoo version (#1032) (@jeremiahpslewis)
- Drop devmode, prepare RL.jl v0.11 for release (#1035) (@jeremiahpslewis)
- Update docs script for new 'limited' RL.jl release (#1038) (@jeremiahpslewis)
- Tabular Approximator fixes (pre v0.11 changes) (#1040) (@jeremiahpslewis)
- Swap RLZoo for RLFarm in CI, drop RLExperiments (#1041) (@jeremiahpslewis)
- Buildkite tweaks for monorepo (#1042) (@jeremiahpslewis)
- Drop archived projects (#1043) (@jeremiahpslewis)
- Simplify Experiment code after dropping RLExperiment (#1044) (@jeremiahpslewis)
- Fix code coverage scope so it ignores test dir (#1045) (@jeremiahpslewis)
- Fix reset and stop conditions (#1046) (@jeremiahpslewis)
- Drop Functors and use Flux.@layer (#1048) (@jeremiahpslewis)
- Fix naming consistency and add missing hook tests (#1049) (@jeremiahpslewis)
- Add SARS tdlearning back to lib (#1050) (@jeremiahpslewis)
- Update FluxModelApproximator references to FluxApproximator (#1051) (@jeremiahpslewis)
- Epsilon Speedy Explorer (#1052) (@jeremiahpslewis)
- Add TotalRewardPerEpisodeLastN hook (#1053) (@jeremiahpslewis)
- Fix abstract_learner for multiplayer games (#1054) (@jeremiahpslewis)
- Update versions (#1055) (@jeremiahpslewis)
- Update Docs for v0.11 release (#1056) (@jeremiahpslewis)
- Update Katex version, fix vulnerability (#1058) (@jeremiahpslewis)
Closed issues: