Download Latest Version v0.8.0.zip (11.0 MB)
Email in envelope

Get an email when there's a new version of ChainerRL

Home / v0.4.0
Name Modified Size InfoDownloads / Week
Parent folder
README.md 2018-07-23 3.2 kB
v0.4.0.tar.gz 2018-07-23 6.1 MB
v0.4.0.zip 2018-07-23 6.3 MB
Totals: 3 Items   12.4 MB 0

Important enhancements

  • TRPO (trust region policy optimization) is added: chainerrl.agents.TRPO.
  • C51 (categorical DQN) is added: chainerrl.agents.CategoricalDQN.
  • NoisyNet is added: chainerrl.links.FactorizedNoisyLinear and chainerrl.links.to_factorized_noisy.
  • Python 3.7 is supported
  • Examples were improved in terms of logging and random seed setting

Important destructive changes

  • The async module is renamed async_ for Python 3.7 support.

All updates

Enhancements

  • TRPO agent (#204)
  • Use numpy random (#206)
  • Add gpus argument for chainerrl.misc.set_random_seed (#207)
  • More check on nesting AttributeSavingMixin (#208)
  • show error message (#210, thanks @corochann!)
  • Add an option to set whether the agent is saved every time the score is improved (#213)
  • Make tests check exit status of subprocesses (#215)
  • make ReplayBuffer.load() compatible with v0.2.0. (#216, thanks @mr4msm!)
  • Add requirements-dev.txt (#222)
  • Align act and act_and_train's signature to the Agent interface (#230, thanks @lyx-x!)
  • Support dtype arg of spaces.Box (#231)
  • Set outdir to results and add help strings (#248)
  • Categorical DQN (C51) (#249)
  • Remove DiscreteActionValue.sample_epsilon_greedy_actions (#259)
  • Remove DQN.compute_q_values (#260)
  • Enable to change batch_states in PPO (#261, thanks @kuni-kuni!)
  • Remove unnecessary declaration and substitution of 'done' in the train_agent function (#271, thanks @uidilr!)

Documentation

  • Update the contribution guide to use pytest (#220)
  • Add docstring to ALE and fix seed range (#234)
  • Fix docstrings of DDPG (#241)
  • Update the algorithm section of README (#246)
  • Add CategoricalDQN to README (#252)
  • Remove unnecessary comments from examples/gym/train_categorical_dqn_gym.py (#255)
  • Update README.md of examples/ale (#275)

Examples

  • Fix OMP_NUM_THREADS setting (#235)
  • Improve random seed setting in ALE examples (#239)
  • Improve random seed setting for all examples (#243)
  • Use gym and atari wrappers instead of chainerrl.envs.ale (#253)
  • Remove unused args from examples/ale/train_categorical_dqn_ale.py and examples/ale/train_dqn_ale.py (#256)
  • Remove unused --profile argument (#258)
  • Hyperlink DOI against preferred resolver (#266, thanks @katrinleinweber!)

Testing

  • Fix import chainer.testing.condition (#200)
  • Use pytest (#209)
  • Fix PCL tests (#211)
  • Test loading v0.2.0 replay buffers (#217)
  • Use assertRaises instead of expectedFailure (#218)
  • Improve travis script (#242)
  • Run autopep8 in travis ci (#247)
  • Switch autopep8 and hacking (#257)
  • Use hacking 1.0 (#262)
  • Fix a too long line of PPO (#264)
  • Update to hacking 1.1.0 (#274)
  • Add tests of DQN's loss functions (#279)

Bugfixes

  • gym 0.9.6 is not working with python2 (#226)
  • Tiny fix: argument passing in SoftmaxDistribution (#228, thanks @lyx-x!)
  • Add docstring to ALE and fix seed range (#234)
  • except both Exception and KeyboardInterrupt (#250, thanks @uenoku!)
  • Switch autopep8 and hacking (#257)
  • Modify async to async_ to support Python 3.7 (#286, thanks @mmilk1231!)
  • Noisy network fixes (#287, thanks @seann999!)
Source: README.md, updated 2018-07-23