Nadav Cohen

 

 

 

 

 

 

Talk: Implicit Biases of Gradient Descent in Offline System Identification and Optimal Control

When learning to control a critical system (e.g., in healthcare or manufacturing), trial and error is often prohibitively dangerous and/or costly. A natural alternative approach is offline system identification and optimal control: using pre-recorded data for offline learning of a system model, and then using the system model for offline learning of an optimal controller. When implemented with overparameterized models (e.g., neural networks) trained via gradient descent (GD), this approach achieves remarkable success. For example, it enables reducing CO2 emissions of industrial manufacturing plants by up to 20%. This success is driven by implicit biases of GD, which yield not only in-distribution generalization, but also out-of-distribution generalization. Towards elucidating this phenomenon, I will present a series of works that theoretically analyze implicit biases of GD when applied to overparameterized linear models in offline system identification and optimal control. The results I will present offer theoretical explanations for the success of GD in controlling critical systems, and suggest potential avenues for enhancing this success. 

Workshop Home Page

Return to the TAU-UIUC Workshop Home Page

Home→