We systematically investigate these challenges and propose novel numerical methods and architectural solutions that mitigate them, making optimization layers more efficient and effective within deep learning pipelines. Our contributions include methods for enhancing computational efficiency by exploiting the iterative nature of optimization problems, tackling issues of gradient bias and variance in high dimensional problems by exploiting parallelism and network learnt priors about the system, improving sample efficiency in reinforcement learning using approximate simulators, and mitigating representational problems with using complicated constrained optimization layers by creating a tight feedback loop between the optimizer state and the network outputs in domains like robotic control and mechanism design with LLMs. We demonstrate these contributions across different applications, ranging from input-optimization problems, 3D pose estimation and reconstruction, differentiable model predictive control and reinforcement learning problems. We also present a new approach for visual-inertial navigation in nanosatellites, highlighting the practical benefits of integrating optimization layers in challenging real-world scenarios.
Thesis Committee Members:
Zico Kolter, Co-chair
Zac Manchester, Co-chair
Geoffrey Gordon
Max Simchowitz
Vladlen Koltun, Apple
