Proximal Policy Optimization
Paper:
https://openai.com/blog/openai-baselines-ppo/
YouTube Video:
https://www.youtube.com/watch?v=5P7I-xPq8u8&list=PLLO4N3-FoY3feUsA3_XZvn5sXy9Ms8ayE&index=2
#reinforcement_learning #optimization
  
  Paper:
https://openai.com/blog/openai-baselines-ppo/
YouTube Video:
https://www.youtube.com/watch?v=5P7I-xPq8u8&list=PLLO4N3-FoY3feUsA3_XZvn5sXy9Ms8ayE&index=2
#reinforcement_learning #optimization
Openai
  
  Proximal Policy Optimization
  We’re releasing a new class of reinforcement learning algorithms, Proximal Policy Optimization (PPO), which perform comparably or better than state-of-the-art approaches while being much simpler to implement and tune. PPO has become the default reinforcement…
  Model Predictive Control: Powerful Optimization Strategy for Feedback Control
https://www.youtube.com/watch?v=YwodGM2eoy4
#optimization
  
  https://www.youtube.com/watch?v=YwodGM2eoy4
#optimization
YouTube
  
  Model Predictive Control
  This lecture provides an overview of model predictive control (MPC), which is one of the most powerful and general control frameworks.  MPC is used extensively in industrial control settings, and can be used with nonlinear systems and systems with constraints…
  An overview of gradient descent optimization algorithms
Abstract: Gradient descent optimization algorithms, while increasingly popular, are often used as black-box optimizers, as practical explanations of their strengths and weaknesses are hard to come by. This article aims to provide the reader with intuitions with regard to the behaviour of different algorithms that will allow her to put them to use. In the course of this overview, we look at different variants of gradient descent, summarize challenges, introduce the most common optimization algorithms, review architectures in a parallel and distributed setting, and investigate additional strategies for optimizing gradient descent
https://arxiv.org/pdf/1609.04747.pdf
#deep_learning #optimization
  Abstract: Gradient descent optimization algorithms, while increasingly popular, are often used as black-box optimizers, as practical explanations of their strengths and weaknesses are hard to come by. This article aims to provide the reader with intuitions with regard to the behaviour of different algorithms that will allow her to put them to use. In the course of this overview, we look at different variants of gradient descent, summarize challenges, introduce the most common optimization algorithms, review architectures in a parallel and distributed setting, and investigate additional strategies for optimizing gradient descent
https://arxiv.org/pdf/1609.04747.pdf
#deep_learning #optimization