Proximal Policy Optimization
Paper:
https://openai.com/blog/openai-baselines-ppo/
YouTube Video:
https://www.youtube.com/watch?v=5P7I-xPq8u8&list=PLLO4N3-FoY3feUsA3_XZvn5sXy9Ms8ayE&index=2
#reinforcement_learning #optimization
Paper:
https://openai.com/blog/openai-baselines-ppo/
YouTube Video:
https://www.youtube.com/watch?v=5P7I-xPq8u8&list=PLLO4N3-FoY3feUsA3_XZvn5sXy9Ms8ayE&index=2
#reinforcement_learning #optimization
Openai
Proximal Policy Optimization
We’re releasing a new class of reinforcement learning algorithms, Proximal Policy Optimization (PPO), which perform comparably or better than state-of-the-art approaches while being much simpler to implement and tune. PPO has become the default reinforcement…
Model Predictive Control: Powerful Optimization Strategy for Feedback Control
https://www.youtube.com/watch?v=YwodGM2eoy4
#optimization
https://www.youtube.com/watch?v=YwodGM2eoy4
#optimization
YouTube
Model Predictive Control
This lecture provides an overview of model predictive control (MPC), which is one of the most powerful and general control frameworks. MPC is used extensively in industrial control settings, and can be used with nonlinear systems and systems with constraints…
An overview of gradient descent optimization algorithms
Abstract: Gradient descent optimization algorithms, while increasingly popular, are often used as black-box optimizers, as practical explanations of their strengths and weaknesses are hard to come by. This article aims to provide the reader with intuitions with regard to the behaviour of different algorithms that will allow her to put them to use. In the course of this overview, we look at different variants of gradient descent, summarize challenges, introduce the most common optimization algorithms, review architectures in a parallel and distributed setting, and investigate additional strategies for optimizing gradient descent
https://arxiv.org/pdf/1609.04747.pdf
#deep_learning #optimization
Abstract: Gradient descent optimization algorithms, while increasingly popular, are often used as black-box optimizers, as practical explanations of their strengths and weaknesses are hard to come by. This article aims to provide the reader with intuitions with regard to the behaviour of different algorithms that will allow her to put them to use. In the course of this overview, we look at different variants of gradient descent, summarize challenges, introduce the most common optimization algorithms, review architectures in a parallel and distributed setting, and investigate additional strategies for optimizing gradient descent
https://arxiv.org/pdf/1609.04747.pdf
#deep_learning #optimization