The OpenAI Gym can be paralleled by the bathEnv.py, which makes the training faster.
You can use the following command to choose which DQN to use:
python main.py --is_double 1 --is_duel 1 --is_per 1 --is_distributional 1 --is_noisy 1 --num_step 3
The output looks like:
Number_of_frame mean_max_Q average_reward variance_reward
https://arxiv.org/abs/1710.02298
https://web.stanford.edu/class/psych209/Readings/MnihEtAlHassibis15NatureControlDeepRL.pdf
https://arxiv.org/abs/1509.06461
https://arxiv.org/abs/1511.06581
DQN, Double DQN, and Duel DQN parts are implemented by Ben and me. https://github.com/bparr/10703
https://arxiv.org/abs/1511.05952
https://arxiv.org/abs/1707.06887
The best and most concise implement of Distributional RL loss function in Tensorflow by far in the world!!! Better and faster than all others' implements that I can find.
https://arxiv.org/pdf/1706.10295.pdf
Thanks to: https://github.com/andrewliao11/NoisyNet-DQN/blob/master/tf_util.py