by Alexander Ludwig & Sören Viegener
This is an implementation of the Rainbow reinforcement learning agent presented by Hessel et al. The implementation uses parallel asynchronous environments and has some extensions to the original Rainbow agent:
- Different neural network architectures, namely the original DQN architecture, the Impala CNN, and D2RL
- Different exploration strategies, namely epsilon-greedy, noisy nets, softmax exploration, and random network distillation
To train the agent with the original Rainbow settings: (note that this requires a LOT of RAM. At least around 50 GB)
python main.py --log_wandb=False
Some of the best episodes of the five games played by setup 3 can be watched on Youtube:
https://youtube.com/playlist?list=PLdeppp6CMwaRKorJJUJzSIcHffu37su-r
This was created as part of the Project Deep Reinforcement Learning at Ulm University