• Stars
    star
    835
  • Rank 54,605 (Top 2 %)
  • Language
  • Created about 9 years ago
  • Updated over 8 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

A list of papers and resources dedicated to deep reinforcement learning

Deep Reinforcement Learning Papers

A list of papers and resources dedicated to deep reinforcement learning.

Please note that this list is currently work-in-progress and far from complete.

TODOs

  • Add more and more papers
  • Improve the way of classifying papers (tags may be useful)
  • Create a policy of this list: curated or comprehensive, how to define "deep reinforcement learning", etc.

Contributing

If you want to inform the maintainer of a new paper, feel free to contact @mooopan. Issues and PRs are also welcome.

Table of Contents

Papers

Deep Value Function

  • S. Lange and M. Riedmiller, Deep Learning of Visual Control Policies, ESANN, 2010. pdf
    • Deep Fitted Q-Iteration (DFQ)
  • V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonglou, D. Wierstra, and M. Riedmiller, Playing Atari with Deep Reinforcement Learning, NIPS 2013 Deep Learning Workshop, 2013. pdf
    • Deep Q-Network (DQN) with experience replay
  • V. Mnih, K. Kavukcuoglu, D. Silver, A. a Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski, S. Petersen, C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran, D. Wierstra, S. Legg, and D. Hassabis, Human-level control through deep reinforcement learning, Nature, 2015. pdf code
    • Deep Q-Network (DQN) with experience replay and target network
  • T. Schaul, D. Horgan, K. Gregor, and D. Silver, Universal Value Function Approximators, ICML, 2015. pdf
  • A. Nair, P. Srinivasan, S. Blackwell, C. Alcicek, R. Fearon, A. De Maria, M. Suleyman, C. Beattie, S. Petersen, S. Legg, V. Mnih, and D. Silver, Massively Parallel Methods for Deep Reinforcement Learning, ICML Deep Learning Workshop, 2015. pdf
    • Gorila (General Reinforcement Learning Architecture)
  • K. Narasimhan, T. Kulkarni, and R. Barzilay, Language Understanding for Text-based Games Using Deep Reinforcement Learning, EMNLP, 2015. pdf supplementary code
    • LSTM-DQN
  • M. Hausknecht and P. Stone, Deep Recurrent Q-Learning for Partially Observable MDPs, arXiv, 2015. arXiv code
  • M. Lai, Giraffe: Using Deep Reinforcement Learning to Play Chess, arXiv. 2015. arXiv code
  • H. van Hasselt, A. Guez, and D. Silver, Deep reinforcement learning with double q-learning, arXiv, 2015. arXiv
    • Double DQN
  • F. Zhang, J. Leitner, M. Milford, B. Upcroft, and P. Corke, Towards Vision-Based Deep Reinforcement Learning for Robotic Motion Control, in ACRA, 2015. pdf
  • T. Schaul, J. Quan, I. Antonoglou, and D. Silver, Prioritized Experience Replay, arXiv, 2015. arXiv
  • Z. Wang, N. de Freitas, and M. Lanctot, Dueling Network Architectures for Deep Reinforcement Learning, arXiv, 2015. arXiv
  • V. François-Lavet, R. Fonteneau, and D. Ernst, How to Discount Deep Reinforcement Learning: Towards New Dynamic Strategies, NIPS Deep Reinforcement Learning Workshop, 2015. arXiv
  • I. Sorokin, A. Seleznev, M. Pavlov, A. Fedorov, and A. Ignateva, Deep Attention Recurrent Q-Network, NIPS Deep Reinforcement Learning Workshop, 2015. arXiv
  • A. A. Rusu, S. G. Colmenarejo, C. Gulcehre, G. Desjardins, J. Kirkpatrick, R. Pascanu, V. Mnih, K. Kavukcuoglu, and R. Hadsell, Policy Distillation, arXiv, 2015. arXiv
  • M. G. Bellemare, G. Ostrovski, A. Guez, P. S. Thomas, and R. Munos, Increasing the Action Gap: New Operators for Reinforcement Learning, AAAI, 2016. arXiv
  • D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot, S. Dieleman, D. Grewe, J. Nham, N. Kalchbrenner, I. Sutskever, T. Lillicrap, M. Leach, K. Kavukcuoglu, T. Graepel, and D. Hassabis, Mastering the game of Go with deep neural networks and tree search, Nature, 2016. pdf
  • T. Zahavy, N. Ben Zrihem, and S. Mannor, Graying the black box: Understanding DQNs, arXiv, 2016. arXiv
  • J. N. Foerster, Y. M. Assael, N. de Freitas, and S. Whiteson, Learning to Communicate to Solve Riddles with Deep Distributed Recurrent Q-Networks, arXiv, 2016. arXiv
  • I. Osband, C. Blundell, A. Pritzel, and B. Van Roy, Deep Exploration via Bootstrapped DQN, arXiv, 2016. arXiv
  • T. Salimans and D. P. Kingma, Weight Normalization : A Simple Reparameterization to Accelerate Training of Deep Neural Networks, arXiv, 2016. arXiv
  • S. Gu, T. Lillicrap, I. Sutskever, and S. Levine, Continuous Deep Q-Learning with Model-based Acceleration, arXiv, 2016. arXiv
  • J. Heinrich and D. Silver, Deep Reinforcement Learning from Self-Play in Imperfect-Information Games, arXiv, 2016. arXiv
  • T. D. Kulkarni, K. R. Narasimhan, A. Saeedi, and J. B. Tenenbaum, Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation, arXiv, 2016. arXiv
  • J. Oh, V. Chockalingam, S. Singh, and H. Lee, Control of Memory, Active Perception, and Action in Minecraft, ICML, 2016. arXiv
  • T. D. Kulkarni, A. Saeedi, S. Gautam, and S. J. Gershman, Deep Successor Reinforcement Learning, arXiv, 2016. arXiv

Deep Policy

  • S. Levine, C. Finn, T. Darrell, and P. Abbeel, End-to-End Training of Deep Visuomotor Policies, arXiv, 2015. arXiv
    • partially observed guided policy search
  • J. Schulman, S. Levine, P. Moritz, M. Jordan, and P. Abbeel, Trust Region Policy Optimization, ICML, 2015. pdf
  • T. Zhang, G. Kahn, S. Levine, and P. Abbeel, Learning Deep Control Policies for Autonomous Aerial Vehicles with MPC-Guided Policy Search, ICRA, 2016. arXiv

Deep Actor-Critic

  • J. Schulman, P. Moritz, S. Levine, M. Jordan, and P. Abbeel, High-Dimensional Continuous Control Using Generalized Advantage Estimation, arXiv, 2015. arXiv
  • T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra, Continuous control with deep reinforcement learning, arXiv, 2015. arXiv
  • D. Balduzzi and M. Ghifary, Compatible Value Gradients for Reinforcement Learning of Continuous Deep Policies, arXiv, 2015. arXiv
  • M. Hausknecht and P. Stone, Deep Reinforcement Learning in Parameterized Action Space, arXiv. 2015. arXiv
  • N. Heess, J. J. Hunt, T. P. Lillicrap, and D. Silver, Memory-based control with recurrent neural networks, NIPS Deep Reinforcement Learning Workshop, 2015. arXiv
  • V. Mnih, A. P. Badia, M. Mirza, A. Graves, T. P. Lillicrap, T. Harley, D. Silver, and K. Kavukcuoglu, Asynchronous Methods for Deep Reinforcement Learning, arXiv, 2016. arXiv

Deep Model

  • B. C. Stadie, S. Levine, and P. Abbeel, Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models, arXiv, 2015. arXiv
  • J. Oh, X. Guo, H. Lee, R. Lewis, and S. Singh, Action-Conditional Video Prediction using Deep Networks in Atari Games, NIPS, 2015. arXiv
  • J. M. Assael, W. Om, T. B. Schön, and M. P. Deisenroth, Data-Efficient Learning of Feedback Policies from Image Pixels using Deep Dynamical Models, arXiv, 2015 arXiv
  • N. Heess, G. Wayne, D. Silver, T. Lillicrap, Y. Tassa, and T. Erez, Learning Continuous Control Policies by Stochastic Value Gradients, NIPS, 2015. arXiv video
  • J. Schmidhuber, On Learning to Think: Algorithmic Information Theory for Novel Combinations of Reinforcement Learning Controllers and Recurrent Neural World Models, arXiv, 2015. arXiv
  • K. Fragkiadaki, P. Agrawal, S. Levine, and J. Malik, Learning Visual Predictive Models of Physics for Playing Billiards, ICLR, 2016. arXiv

Application to Non-RL Tasks

  • J. C. Caicedo and S. Lazebnik, Active Object Localization with Deep Reinforcement Learning, ICCV, 2015. pdf
  • H. Guo, Generating Text with Deep Reinforcement Learning, arXiv, 2015. arXiv
  • S. Hansen, Using Deep Q-Learning to Control Optimization Hyperparameters, arXiv, 2016. arXiv

Unclassified

  • X. Guo, S. Singh, H. Lee, R. Lewis, and X. Wang, Deep learning for real-time Atari game play using offline Monte-Carlo tree search planning, NIPS, 2014. pdf video
  • S. Mohamed and D. J. Rezende, Variational Information Maximisation for Intrinsically Motivated Reinforcement Learning, arXiv, 2015. arXiv
  • Y. Liang, M. C. Machado, E. Talvitie, and M. Bowling, State of the Art Control of Atari Games Using Shallow Reinforcement Learning, arXiv, 2015. arXiv
  • A. Tamar, S. Levine, and P. Abbeel, Value Iteration Networks, arXiv, 2016. arXiv
  • C. Blundell, G. Deepmind, B. Uria, A. Pritzel, Y. Li, A. Ruderman, J. Z. Leibo, J. Rae, D. Wierstra, and D. Hassabis, Model-Free Episodic Control, arXiv, 2016. arXiv

Talks/Slides

  • S. Levine, Deep Learning for Decision Making and Control, 2015. video
  • D. Silver, Deep Reinforcement Learning, ICLR, 2015. video1 video2 slides
  • D. Silver, Deep Reinforcement Learning, UAI, 2015. video

Miscellaneous

More Repositories

1

async-rl

Replicating "Asynchronous Methods for Deep Reinforcement Learning" (http://arxiv.org/abs/1602.01783)
Python
400
star
2

dqn-in-the-caffe

An implementation of Deep Q-Network using Caffe
C++
213
star
3

deep-ensemble-uncertainty

An implementation of "Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles" (http://arxiv.org/abs/1612.01474)
Jupyter Notebook
34
star
4

predictron

WIP implementation of "The Predictron: End-To-End Learning and Planning" (http://arxiv.org/abs/1612.08810) in Chainer
Python
11
star
5

chainer-cocob

COCOB-Backprop (https://arxiv.org/abs/1705.07795) implementation for Chainer
Python
6
star
6

SexprParser

This is S-expression parser in C++11. It is aimed at being used in General Game Playing.
C++
3
star
7

dobutsushogi

Dobutsu Shogi in Game Description Language (GDL v1)
2
star
8

ggpe

General Game Playing Engine in C++11 using YAP Prolog
C++
2
star
9

chainer-weight-normalization

Weight normalization https://arxiv.org/abs/1602.07868
Python
1
star
10

chainer-eve

An Eve optimizer implementation in Chainer
Python
1
star
11

mcts

Monte-Carlo Tree Search Implementation for general purpose.
C++
1
star
12

chainer-elu

Chainer implementation of Exponential Linear Unit (ELU)
Python
1
star
13

nonogram

JavaScript
1
star
14

chainer-oplu

Orthogonal Permuatation Linear Unit (OPLU) https://arxiv.org/abs/1604.02313v3
Python
1
star
15

chainer-entropy-adam

Chainer-based implementation of Entropy-Adam https://arxiv.org/abs/1611.01838
Python
1
star
16

project_euler

Project Euler solutions written by muupan. Most of the solutions are written in ruby or python.
Python
1
star