Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
README.md | 2018-07-23 | 3.2 kB | |
v0.4.0.tar.gz | 2018-07-23 | 6.1 MB | |
v0.4.0.zip | 2018-07-23 | 6.3 MB | |
Totals: 3 Items | 12.4 MB | 0 |
Important enhancements
- TRPO (trust region policy optimization) is added:
chainerrl.agents.TRPO
. - C51 (categorical DQN) is added:
chainerrl.agents.CategoricalDQN
. - NoisyNet is added:
chainerrl.links.FactorizedNoisyLinear
andchainerrl.links.to_factorized_noisy
. - Python 3.7 is supported
- Examples were improved in terms of logging and random seed setting
Important destructive changes
- The
async
module is renamedasync_
for Python 3.7 support.
All updates
Enhancements
- TRPO agent (#204)
- Use numpy random (#206)
- Add gpus argument for chainerrl.misc.set_random_seed (#207)
- More check on nesting AttributeSavingMixin (#208)
- show error message (#210, thanks @corochann!)
- Add an option to set whether the agent is saved every time the score is improved (#213)
- Make tests check exit status of subprocesses (#215)
- make ReplayBuffer.load() compatible with v0.2.0. (#216, thanks @mr4msm!)
- Add requirements-dev.txt (#222)
- Align act and act_and_train's signature to the Agent interface (#230, thanks @lyx-x!)
- Support dtype arg of spaces.Box (#231)
- Set outdir to results and add help strings (#248)
- Categorical DQN (C51) (#249)
- Remove DiscreteActionValue.sample_epsilon_greedy_actions (#259)
- Remove DQN.compute_q_values (#260)
- Enable to change batch_states in PPO (#261, thanks @kuni-kuni!)
- Remove unnecessary declaration and substitution of 'done' in the train_agent function (#271, thanks @uidilr!)
Documentation
- Update the contribution guide to use pytest (#220)
- Add docstring to ALE and fix seed range (#234)
- Fix docstrings of DDPG (#241)
- Update the algorithm section of README (#246)
- Add CategoricalDQN to README (#252)
- Remove unnecessary comments from examples/gym/train_categorical_dqn_gym.py (#255)
- Update README.md of examples/ale (#275)
Examples
- Fix OMP_NUM_THREADS setting (#235)
- Improve random seed setting in ALE examples (#239)
- Improve random seed setting for all examples (#243)
- Use gym and atari wrappers instead of chainerrl.envs.ale (#253)
- Remove unused args from examples/ale/train_categorical_dqn_ale.py and examples/ale/train_dqn_ale.py (#256)
- Remove unused --profile argument (#258)
- Hyperlink DOI against preferred resolver (#266, thanks @katrinleinweber!)
Testing
- Fix import chainer.testing.condition (#200)
- Use pytest (#209)
- Fix PCL tests (#211)
- Test loading v0.2.0 replay buffers (#217)
- Use assertRaises instead of expectedFailure (#218)
- Improve travis script (#242)
- Run autopep8 in travis ci (#247)
- Switch autopep8 and hacking (#257)
- Use hacking 1.0 (#262)
- Fix a too long line of PPO (#264)
- Update to hacking 1.1.0 (#274)
- Add tests of DQN's loss functions (#279)
Bugfixes
- gym 0.9.6 is not working with python2 (#226)
- Tiny fix: argument passing in SoftmaxDistribution (#228, thanks @lyx-x!)
- Add docstring to ALE and fix seed range (#234)
- except both Exception and KeyboardInterrupt (#250, thanks @uenoku!)
- Switch autopep8 and hacking (#257)
- Modify
async
toasync_
to support Python 3.7 (#286, thanks @mmilk1231!) - Noisy network fixes (#287, thanks @seann999!)