Speaker
Description
Model Predictive Control (MPC) and Reinforcement Learning (RL) are two of the most prominent methods for computing control laws based on optimization. In both cases, the resulting controllers approximate infinite-horizon optimal controllers, where the objective of the optimization may range from stabilization of a set-point to energy efficiency to yield maximization. However, for both methods the computational effort may make their application infeasible for large-scale or complex problems. In this talk we explain the basic functioning of both methods and then present situations in which the methods provably work well, by identifying beneficial structures of the solutions of optimal control problems. In the case of MPC we focus on the so-called turnpike property of optimal trajectories, while for Deep RL (i.e., RL with deep neural networks as approximators) we look at the compositional structure of optimal value functions. Examples and use cases from academia and from industry illustrate the theoretical findings.