• Stars
    star
    711
  • Rank 63,679 (Top 2 %)
  • Language
    Jupyter Notebook
  • Created almost 6 years ago
  • Updated almost 6 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Reinforcement Learning Tutorial with Demo: DP (Policy and Value Iteration), Monte Carlo, TD Learning (SARSA, QLearning), Function Approximation, Policy Gradient, DQN, Imitation, Meta Learning, Papers, Courses, etc..

Reinforcement Learning (RL) Tutorial

There are many RL tutorials, courses, papers in the internet. This one summarizes all of the RL tutorials, RL courses, and some of the important RL papers including sample code of RL algorithms. It will continue to be updated over time.

Keywords: Dynamic Programming (Policy and Value Iteration), Monte Carlo, Temporal Difference (SARSA, QLearning), Approximation, Policy Gradient, DQN, Imitation Learning, Meta-Learning, RL papers, RL courses, etc.

NOTE: This tutorial is only for education purpose. It is not academic study/paper. All related references are listed at the end of the file.

Table of Contents

What is RL?

Machine learning mainly consists of three methods: Supervised Learning, Unsupervised Learning and Reinforcement Learning. Supervised Learning provides mapping functionality between input and output using labelled dataset. Some of the supervised learning methods: Linear Regression, Support Vector Machines, Neural Networks, etc. Unsupervised Learning provides grouping and clustering functionality. Some of the unsupervised learning methods: K-Means, DBScan, etc. Reinforcement Learning is different from supervised and unsupervised learning. RL provides behaviour learning.

"A reinforcement learning algorithm, or agent, learns by interacting with its environment. The agent receives rewards by performing correctly and penalties for performing incorrectly. The agent learns without intervention from a human by maximizing its reward and minimizing its penalty" *. RL agents are used in different applications: Robotics, self driving cars, playing atari games, managing investment portfolio, control problems. I am believing that like many AI laboratories do, reinforcement learning with deep learning will be a core technology in the future.

rl-agent-env [Sutton & Barto Book: RL: An Introduction]

Markov Decision Process :

  • It consists of five tuples: status, actions, rewards, state transition probability, discount factor.
  • Markov decision processes formally describe an environment for reinforcement learning.
  • There are 3 techniques for solving MDPs: Dynamic Programming (DP) Learning, Monte Carlo (MC) Learning, Temporal Difference (TD) Learning.

mdps [David Silver Lecture Notes]

Markov Property :

A state St is Markov if and only if P[St+1|St] =P[St+1|S1,...,St]

RL Components :

Rewards :

  • A reward Rt is a scalar feedback signal.
  • The agent’s job is to maximise cumulative reward

State Transition Probability :

  • p(s′,r|s,a).= Pr{St=s′,Rt=r|St-1=s,At−1=a},

Discount Factor :

  • The discount γ∈[0,1] is the present value of future rewards.

Return :

  • The return Gt is the total discounted reward from time-step t.

return [David Silver Lecture Notes]

Value Function :

  • Value function is a prediction of future reward. How good is each state and/or action.
  • The value function v(s) gives the long-term value of state s
  • Vπ(s) =Eπ[Rt+1+γRt+22Rt+3+...|St=s]
  • Value function has two parts: immediate reward and discounted value of successor state.

value_function [David Silver Lecture Notes]

Policy (π) :

A policy is the agent’s behaviour. It is a map from state to action.

  • Deterministic policy: a=π(s).
  • Stochastic policy: π(a|s) =P[At=a|St=s].

State-Value Function :

state-value [David Silver Lecture Notes]

Action-Value Function :

action- value [David Silver Lecture Notes]

Optimal Value Functions :

optimal-value [David Silver Lecture Notes]

Planning vs RL :

Planning:

  • Rules of the game are known.
  • A model of the environment is known.
  • The agent performs computations with its mode.
  • The agent improves its policy.

RL:

  • The environment is initially unknown.
  • The agent interacts with the environment.
  • The agent improves its policy.

Exploration and Exploitation :

  • Reinforcement learning is like trial-and-error learning.
  • The agent should discover a good policy.
  • Exploration finds more information about the environment (Gather more information).
  • Exploitation exploits known information to maximise reward (Make the best decision given current information).

Prediction & Control Problem (Pattern of RL algorithms) :

  • Prediction: evaluate the future (Finding value given a policy).
  • Control: optimise the future (Finding optimal/best policy).

Grid World :

  • Grid World is a game for demonstration. 12 positions, 11 states, 4 actions. Our aim is to find optimal policy.
  • Demo Code: gridWorldGame.py

grid-world

optimal-policy-grid

Dynamic Programming Method (DP): Full Model :

  • Dynamic Programming is a very general solution method for problems which have two properties: 1.Optimal substructure, 2.Overlapping subproblems.
  • Markov decision processes satisfy both properties. Bellman equation gives recursive decomposition. Value function stores and reuses solutions.
  • In DP method, full model is known, It is used for planning in an MDP.
  • There are 2 methods: Policy Iteration, Value Iteration.
  • DP uses full-width backups.
  • DP is effective for medium-sized problems (millions of states).
  • For large problems DP suffers Bellman’s curse of dimensionality.
  • Disadvantage of DP: requires full model of environment, never learns from experience.

Policy Iteration (with Pseudocode) :

policy iteration [David Silver Lecture Notes]

policy-iteration

Policy Evaluation (with Pseudocode) :

  • Problem: evaluate a given policy π.
  • Solution: iterative application of Bellman expectation backup.
  • v1 → v2→ ... → vπ.

iterative-policy-evaluation

Policy Improvement :

policy-improvement

Value Iteration (with Pseudocode) :

  • Policy iteration has 2 inner loop. However, value iteration has a better solution.
  • It combines policy evaluation and policy improvement into one step.
  • Problem: find optimal policy π.
  • Solution: iterative application of Bellman optimality backup.

value-iteration

Monte Carlo (MC) Method :

  • Demo Code: monte_carlo_demo.ipynb
  • MC methods learn directly from episodes of experience.
  • MC is model-free : no knowledge of MDP transitions / rewards.
  • MC uses the simplest possible idea: value = mean return.
  • Episode must terminate before calculating return.
  • Average return is calculated instead of using true return G.
  • First Visit MC: The first time-step t that state s is visited in an episode.
  • Every Visit MC: Every time-step t that state s is visited in an episode.

MC Calculating Returns (with Pseudocode) :

mc-calculating-returns

First-Visit MC (with Pseudocode) (MC Prediction Problem) :

first-visit-mc

MC Exploring-Starts (with Pseudocode) (MC Control Problem) :

  • Demo Code: monte_carlo_es_demo.ipynb
  • State s and Action a is randomly selected for all starting points.
  • Use Q instead of V
  • Update the policy after every episode, keep updating the same Q in-place.

mc-control1

MC Epsilon Greedy (without Exploring Starts) :

  • Demo Code: monte_carlo_epsilon_greedy_demo.ipynb
  • MC Exploring start is infeasible, because in real problems we can not calculate all edge cases (ex: in self-driving car problem, we can not calculate all cases).
  • Randomly selection for all starting points in code is removed.
  • Change policy to sometimes be random.
  • This random policy is Epsilon-Greedy (like multi-armed bandit problem)

Temporal Difference (TD) Learning Method :

  • Demo Code: td0_prediction.ipynb
  • TD methods learn directly from episodes of experience.
  • TD updates a guess towards a guess
  • TD learns from incomplete episodes, by bootstrapping.
  • TD uses bootstrapping like DP, TD learns experience like MC (combines MC and DP).

MC - TD Difference :

  • MC and TD learn from experience.
  • TD can learn before knowing the final outcome.
  • TD can learn online after every step. MC must wait until end of episode before return is known.
  • TD can learn without the final outcome.
  • TD can learn from incomplete sequences. MC can only learn from complete sequences.
  • TD works in continuing environments. MC only works for episodic environments.
  • MC has high variance, zero bias. TD has low variance, some bias.

mc-td-dif1 mc-td-dif2 mc-td-dif3 [David Silver Lecture Notes]

MC - TD - DP Difference in Visual :

mc-td-dp [David Silver Lecture Notes]

SARSA (TD Control Problem, On-Policy) :

  • Demo Code: SARSA_demo.ipynb
  • In on-policy learning the Q(s,a) function is learned from actions, we took using our current policy π.

updatingwithsarsa sarsa-algo [David Silver Lecture Notes]

Q-Learning (TD Control Problem, Off-Policy) :

  • Demo Code: q_learning_demo.ipynb
  • Looks like SARSA, instead of choosing a' based on argmax of Q, Q(s,a) is updated directly with max over Q(s',a')
  • In off-policy learning the Q(s,a) function is learned from different actions (for example, random actions). We even don't need a policy at all.

qfunction

updatingwithqlearning qlearning-algo [David Silver Lecture Notes]

Function Approximation :

  • Demo Code: func_approx_q_learning_demo.ipynb
  • Reinforcement learning can be used to solve large problems, e.g Backgammon: 1020 states; Computer Go: 10170 states; Helicopter: continuous state space.
  • So far we have represented value function by a lookup table, Every state s has an entry V(s) or every state-action pair s, a has an entry Q(s,a).
  • There are too many states and/or actions to store in memory. It is too slow to learn the value of each state individually. Tabulated Q may not fit memory.
  • Solution for large MDPs:

func-appr

  • Differentiable function approximators can be used: Linear combinations of features, Neural Networks.

value-func-appr

func-appr2 [David Silver Lecture Notes]

Feature Vector :

feature-vectors [David Silver Lecture Notes]

Open AI Gym Environment :

gym

Policy-Based Methods :

  • DP, MC and TD Learning methods are value-based methods (Learnt Value Function, Implicit policy).
  • In value-based methods, a policy was generated directly from the value function (e.g. using epsilon-greedy)
  • In policy-based, we will directly parametrise the policy ( πθ(s,a) =P[a|s,θ) ).
  • Policy Gradient method is a policy-based method (No Value Function, Learnt Policy).

policy-based

Advantages:

  • Better convergence properties,
  • Effective in high-dimensional or continuous action spaces,
  • Can learn stochastic policies.

Disadvantages:

  • Typically converge to a local rather than global optimum.
  • Evaluating a policy is typically inefficient and high variance.

Policy Objective Functions :

  • Policy based reinforcement learning is an optimisation problem.
  • Find θ that maximises J(θ).

policy-objective-func [David Silver Lecture Notes]

Policy-Gradient :

policy-gradient

score-function

policy-gradient-theorem [David Silver Lecture Notes]

Monte-Carlo Policy Gradient (REINFORCE) :

reinforce [David Silver Lecture Notes]

Actor-Critic :

  • Actor-Critic method is a policy-based method (Learnt Value Function, Learnt Policy).

actor-critique-intro [David Silver Lecture Notes]

  • The critic is solving a familiar problem: policy evaluation.

Action-Value Actor-Critic :

action-value-actor-critic

advantage_func [David Silver Lecture Notes]

  • The advantage function can significantly reduce variance of policy gradient.
  • So the critic should really estimate the advantage function.

Actor-critic algorithm: A3C :

a3c

Different Policy Gradients :

policy-gradient-summary [David Silver Lecture Notes]

Model-Based RL :

model-based-rl

model-based-rl2 [David Silver Lecture Notes]

  • Favourite planning algorithms: Value iteration,Policy iteration,Tree search,etc..
  • Sample-based Modeling: A simple but powerful approach to planning. Use the model only to generate samples. Sample experience from model.
  • Apply model-free RL to samples, e.g.: Monte-Carlo control, SARSA, Q-Learning.
  • Model-based RL is only as good as the estimated model.
  • When the model is inaccurate, planning process will compute a suboptimal policy: 1.when model is wrong, use model-free RL; 2.reason explicitly about model uncertainty.

Real and Simulated Experience :

real-simulated-exp

dyna-arch [David Silver Lecture Notes]

Dyna-Q Algorithm :

dynaq [David Silver Lecture Notes]

Sim-Based Search :

sim-based-search [David Silver Lecture Notes]

MC-Tree-Search :

  • AlphaGo- Supervised learning + policy gradients + value functions + Monte Carlo tree search D. Silver, A. Huang, C. J.Maddison, A. Guez, L. Sifre, et al. “Mastering the game of Go with deep neural networks and tree search”. Nature (2016).
  • Highly selective best-first search.
  • Evaluates states dynamically (unlike e.g. DP).
  • Uses sampling to break curse of dimensionality.
  • Computationally efficient, anytime, parallelisable.
  • Works for “black-box” models (only requires samples).

mc-tree-search [David Silver Lecture Notes]

Temporal-Difference Search :

  • Simulation-based search.
  • Using TD instead of MC (bootstrapping).
  • MC tree search applies MC control to sub-MDP from now.
  • TD search applies Sarsa to sub-MDP from now.
  • For simulation-based search, bootstrapping is also helpful.
  • TD search is usually more efficient than MC search.
  • TD(λ) search can be much more efficient than MC search.

td-search [David Silver Lecture Notes]

RL in Games :

rl-in-games [David Silver Lecture Notes]

Deep Q Learning (Deep Q-Networks: DQN) :

  • Gradient descent is simple and appealing. But it is not sample efficient.
  • Batch methods seek to find the best fitting value function.
  • Given the agent’s experience (“training data”)

Experience Replay :

dqn-experience-replay

DQN in Atari :

-V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, et al. “Playing Atari with Deep Reinforcement Learning”. (2013)

dqn-in-atari

Imitation Learning :

imitation-learning1

imitation-learning2

Dagger: Dataset Aggregation :

Paper: https://www.cs.cmu.edu/~sross1/publications/Ross-AIStats11-NoRegret.pdf

dagger

PLATO: Policy Learning with Adaptive Trajectory Optimization :

plato

One-Shot Imitation Learning :

Meta-Learning :

  • Meta-learning = learning to learn
  • Supervised meta-learning = supervised learning with datapoints that are entire datasets
  • If we can meta-learn a faster reinforcement learner, we can learn new tasks efficiently!
  • What can a meta-learned learner do differently? 1.Explore more intelligently, 2.Avoid trying actions that are know to be useless, 3.Acquire the right features more quickly.
  • The promise of meta-learning: use past experience to simply acquire a much more efficient deep RL algorithm

meta-learning

POMDPs (Partial Observable MDP) :

pomdps [David Silver Lecture Notes]

Resources :

Important RL Papers :

References :

More Repositories

1

Fast-Kubernetes

This repo covers Kubernetes with LABs: Kubectl, Pod, Deployment, Service, PV, PVC, Rollout, Multicontainer, Daemonset, Taint-Toleration, Job, Ingress, Kubeadm, Helm, etc.
PowerShell
1,773
star
2

LSTM_RNN_Tutorials_with_Demo

LSTM-RNN Tutorial with LSTM and RNN Tutorial with Demo with Demo Projects such as Stock/Bitcoin Time Series Prediction, Sentiment Analysis, Music Generation using Keras-Tensorflow
Python
672
star
3

Fast-Ansible

This repo covers Ansible with LABs: Multipass, Commands, Modules, Playbooks, Tags, Managing Files and Servers, Users, Roles, Handlers, Host Variables, Templates and details.
Jinja
625
star
4

Fast-Docker

This repo covers containerization and Docker Environment: Docker File, Image, Container, Commands, Volumes, Networks, Swarm, Stack, Service, possible scenarios.
Dockerfile
566
star
5

Fast-Pytorch

Pytorch Tutorial, Pytorch with Google Colab, Pytorch Implementations: CNN, RNN, DCGAN, Transfer Learning, Chatbot, Pytorch Sample Codes
Jupyter Notebook
427
star
6

Generative_Models_Tutorial_with_Demo

Generative Models Tutorial with Demo: Bayesian Classifier Sampling, Variational Auto Encoder (VAE), Generative Adversial Networks (GANs), Popular GANs Architectures, Auto-Regressive Models, Important Generative Model Papers, Courses, etc..
Jupyter Notebook
326
star
7

Fast-Terraform

This repo covers Terraform (Infrastructure as Code) with LABs using AWS and AWS Sample Projects: Resources, Variables, Meta Arguments, Provisioners, Dynamic Blocks, Modules, Provisioning AWS Resources (EC2, EBS, EFS, VPC, IAM Policies, Roles, ECS, ECR, Fargate, EKS, Lambda, API-Gateway, ELB, S3, etc.
HCL
274
star
8

CNN-TA

Algorithmic Financial Trading with Deep Convolutional Neural Networks: Time Series to Image Conversion Approach: A novel algorithmic trading model CNN-TA using a 2-D convolutional neural network based on image processing properties.
Java
108
star
9

Fast-Kubeflow

This repo covers Kubeflow Environment with LABs: Kubeflow GUI, Jupyter Notebooks on pods, Kubeflow Pipelines, Experiments, KALE, KATIB (AutoML: Hyperparameter Tuning), KFServe (Model Serving), Training Operators (Distributed Training), Projects, etc.
Python
75
star
10

SparkDeepMlpGADow30

A Deep Neural-Network based (Deep MLP) Stock Trading System based on Evolutionary (Genetic Algorithm) Optimized Technical Analysis Parameters (using Apache Spark MLlib)
Java
64
star
11

TimeSeries2DBarChartImageCNN

Conversion of the time series values to 2-D stock bar chart images and prediction using CNN (using Keras-Tensorflow)
Python
38
star
12

SparkMlpDow30

A new stock trading and prediction model based on a MLP neural network utilizing technical analysis indicator values as features (using Apache Spark MLlib)
Java
35
star
13

AIMap

Map of Artificial Intelligence: Classifications, Approaches, Algorithms, Libraries, Tools, State of Art Studies, Awesome Repos, etc..
21
star
14

PolicyGradient_PongGame

Pong Game problem solving using RL - Policy Gradient with OpenAI Gym Framework and Tensorflow
Python
14
star
15

Qlearning_MountainCar

Mountain Car problem solving using RL - QLearning with OpenAI Gym Framework
Python
8
star
16

IoTSmartHomeOntologySimulator

A smart home sensor ontology (that is a specialized ontology based on the Semantic Sensor Networks (SSN) ontology) and simulation evironment of a smart home use case
Java
5
star
17

SentimentAnalysis

Sentences are classified in 5 different sentiment using LSTM (Keras). Results are expressed with emoji characters.
Python
5
star
18

NeuralStyleTransfer

Art/Painting Generation using AI (Neural Style Transfer) using Tensorflow
Python
5
star
19

BasicLSTM

The aim of this implementation is to help to learn structure of basic LSTM (LSTM cell forward, LSTM cell backward, etc..)
Python
4
star
20

IoTWeatherSensorsAnalysis

Proposed "An Extended IoT Framework" learning part is presented with a use case on weather data clustering analysis. Sensor faults and anomalies are determined using K-means clustering (using scikit-learn)
Python
4
star
21

MusicGeneration

Music generation with LSTM model (Keras)
Python
3
star
22

QLearning_CartPole

Cart Pole problem solving using RL - QLearning with OpenAI Gym Framework
Python
2
star
23

modelbuild_pipeline

Test Repo for Model Build
Python
1
star