Difference between revisions of "Reinforcement Learning (RL)"

From
Jump to: navigation, search
m
m
Line 44: Line 44:
 
* [[World Models]]
 
* [[World Models]]
 
* [[Google DeepMind AlphaGo Zero]]
 
* [[Google DeepMind AlphaGo Zero]]
* [https://venturebeat.com/2019/06/19/googles-ai-picks-which-machine-learning-models-will-produce-the-best-results/ Google’s AI picks which machine learning models will produce the best results | Kyle Wiggers - VentureBeat] off-policy classification,” or OPC, which evaluates the performance of AI-driven agents by treating evaluation as a classification problem
+
* [https://venturebeat.com/2019/06/19/googles-ai-picks-which-machine-learning-models-will-produce-the-best-results/ Google’s AI picks which machine learning models will produce the best results | Kyle Wiggers - VentureBeat] off-policy classification,” or OPC, which evaluates the performance of AI-driven [[Agents|agents]] by treating evaluation as a classification problem
 
* [http://www.amazon.com/Deep-Reinforcement-Learning-Hands-Q-networks/dp/1788834240 Deep Reinforcement Learning Hands-On: Apply modern RL methods, with deep Q-networks, value iteration, policy gradients, TRPO, AlphaGo Zero and more | Maxim Lapan]
 
* [http://www.amazon.com/Deep-Reinforcement-Learning-Hands-Q-networks/dp/1788834240 Deep Reinforcement Learning Hands-On: Apply modern RL methods, with deep Q-networks, value iteration, policy gradients, TRPO, AlphaGo Zero and more | Maxim Lapan]
 
* [http://github.com/Pulkit-Khandelwal/Reinforcement-Learning-Notebooks Reinforcement-Learning-Notebooks] - A collection of Reinforcement Learning algorithms from Sutton and Barto's book and other research papers implemented in Python  
 
* [http://github.com/Pulkit-Khandelwal/Reinforcement-Learning-Notebooks Reinforcement-Learning-Notebooks] - A collection of Reinforcement Learning algorithms from Sutton and Barto's book and other research papers implemented in Python  
Line 62: Line 62:
  
  
This is a bit similar to the traditional type of data analysis; the algorithm discovers through trial and error and decides which action results in greater rewards. Three major components can be identified in reinforcement learning functionality: the agent, the environment, and the actions. The agent is the learner or decision-maker, the environment includes everything that the agent interacts with, and the actions are what the agent can do. Reinforcement learning occurs when the agent chooses actions that maximize the expected reward over a given time. This is best achieved when the agent has a good policy to follow. [http://www.simplilearn.com/what-is-machine-learning-and-why-it-matters-article Machine Learning: What it is and Why it Matters | Priyadharshini @ simplilearn]
+
This is a bit similar to the traditional type of data analysis; the algorithm discovers through trial and error and decides which action results in greater rewards. Three major components can be identified in reinforcement learning functionality: the [[Agents|agent]], the environment, and the actions. The [[Agents|agent]] is the learner or decision-maker, the environment includes everything that the [[Agents|agent]] interacts with, and the actions are what the [[Agents|agent]] can do. Reinforcement learning occurs when the [[Agents|agent]] chooses actions that maximize the expected reward over a given time. This is best achieved when the [[Agents|agent]] has a good policy to follow. [http://www.simplilearn.com/what-is-machine-learning-and-why-it-matters-article Machine Learning: What it is and Why it Matters | Priyadharshini @ simplilearn]
  
 
http://slideplayer.com/24/7469154/big_thumb.jpg
 
http://slideplayer.com/24/7469154/big_thumb.jpg
Line 84: Line 84:
 
<youtube>313kbpBq8Sg</youtube>
 
<youtube>313kbpBq8Sg</youtube>
  
== [http://pythonprogramming.net/search/?q=q+learning Q Learning Algorithm and Agent - Reinforcement Learning  w/ Python Tutorial | Sentdex - Harrison] ==
+
== [http://pythonprogramming.net/search/?q=q+learning Q Learning Algorithm and] [[Agents|Agent]] - Reinforcement Learning  w/ Python Tutorial | Sentdex - Harrison ==
  
 
[http://pythonprogramming.net/q-learning-reinforcement-learning-python-tutorial P.1]
 
[http://pythonprogramming.net/q-learning-reinforcement-learning-python-tutorial P.1]
Line 117: Line 117:
 
⌨️ ([http://www.youtube.com/watch?v=ELE2_Mftqoc&t=3123s 00:52:03]) Deep Q Learning with Pytorch Part 1: The Q Network  
 
⌨️ ([http://www.youtube.com/watch?v=ELE2_Mftqoc&t=3123s 00:52:03]) Deep Q Learning with Pytorch Part 1: The Q Network  
  
⌨️ ([http://www.youtube.com/watch?v=ELE2_Mftqoc&t=3981s 01:06:21]) Deep Q Learning with Pytorch part 2: Coding the Agent  
+
⌨️ ([http://www.youtube.com/watch?v=ELE2_Mftqoc&t=3981s 01:06:21]) Deep Q Learning with Pytorch part 2: Coding the [[Agents|Agent]]
  
 
⌨️ ([http://www.youtube.com/watch?v=ELE2_Mftqoc&t=5334s 01:28:54]) Deep Q Learning with Pytorch part 3
 
⌨️ ([http://www.youtube.com/watch?v=ELE2_Mftqoc&t=5334s 01:28:54]) Deep Q Learning with Pytorch part 3

Revision as of 18:01, 4 February 2023

YouTube search... ...Google search

___________________________________________________________



DeepMind says reinforcement learning is ‘enough’ to reach general AI ... Some scientists believe that assembling multiple narrow AI modules will produce higher intelligent systems.



This is a bit similar to the traditional type of data analysis; the algorithm discovers through trial and error and decides which action results in greater rewards. Three major components can be identified in reinforcement learning functionality: the agent, the environment, and the actions. The agent is the learner or decision-maker, the environment includes everything that the agent interacts with, and the actions are what the agent can do. Reinforcement learning occurs when the agent chooses actions that maximize the expected reward over a given time. This is best achieved when the agent has a good policy to follow. Machine Learning: What it is and Why it Matters | Priyadharshini @ simplilearn

big_thumb.jpg

Machine_Learning_5.jpg

Control-based: When running a Reinforcement Learning (RL) policy in the real world, such as controlling a physical robot on visual inputs, it is non-trivial to properly track states, obtain reward signals or determine whether a goal is achieved for real. The visual data has a lot of noise that is irrelevant to the true state and thus the equivalence of states cannot be inferred from pixel-level comparison. Self-supervised representation learning has shown great potential in learning useful state embedding that can be used directly as input to a control policy.


Q Learning Algorithm and Agent - Reinforcement Learning w/ Python Tutorial | Sentdex - Harrison

P.1

P.2

P.3

P.4

P.5

P.6

Reinforcement Learning | Phil Tabor

Reinforcement learning is an area of machine learning that involves taking right action to maximize reward in a particular situation. In this full tutorial course, you will get a solid foundation in reinforcement learning core topics. The course covers Q learning, State-Action-Reward-State-Action (SARSA), double Q learning, Deep Q Learning (DQN), and Policy Gradient (PG) methods. These algorithms are employed in a number of environments from the open AI gym, including space invaders, breakout, and others. The deep learning portion uses Tensorflow and PyTorch. The course begins with more modern algorithms, such as deep q learning and Policy Gradient (PG) methods, and demonstrates the power of reinforcement learning. Then the course teaches some of the fundamental concepts that power all reinforcement learning algorithms. These are illustrated by coding up some algorithms that predate deep learning, but are still foundational to the cutting edge. These are studied in some of the more traditional environments from the OpenAI Gym, like the cart pole problem.

⌨️ (00:00:00) Introduction

⌨️ (00:01:30) Intro to Deep Q Learning

⌨️ (00:08:56) How to Code Deep Q Learning in Tensorflow

⌨️ (00:52:03) Deep Q Learning with Pytorch Part 1: The Q Network

⌨️ (01:06:21) Deep Q Learning with Pytorch part 2: Coding the Agent

⌨️ (01:28:54) Deep Q Learning with Pytorch part 3

⌨️ (01:46:39) Intro to Policy Gradients 3: Coding the main loop

⌨️ (01:55:01) How to Beat Lunar Lander with Policy Gradients

⌨️ (02:21:32) How to Beat Space Invaders with Policy Gradients

⌨️ (02:34:41) How to Create Your Own Reinforcement Learning Environment Part 1

⌨️ (02:55:39) How to Create Your Own Reinforcement Learning Environment Part 2

⌨️ (03:08:20) Fundamentals of Reinforcement Learning

⌨️ (03:17:09) Markov Decision Processes

⌨️ (03:23:02) The Explore Exploit Dilemma

⌨️ (03:29:19) Reinforcement Learning in the Open AI Gym: SARSA

⌨️ (03:39:56) Reinforcement Learning in the Open AI Gym: Double Q Learning

⌨️ (03:54:07) Conclusion


Jump Start

Lunar Lander: Deep Q learning is Easy in PyTorch

Lunar Lander: How to Beat Lunar Lander with Policy Gradients | Tensorflow Tutorial

Breakout: How to Code Deep Q Learning in Tensorflow (Tutorial)

Gridworld: How To Create Your Own Reinforcement Learning Environments

Reinforcement Learning (RL) from Human Feedback (RLHF)