Computer Science Thesis Oral

Thursday, May 2, 2019 - 11:00am


4305 Newell-Simon Hall



Differentiable Optimization-Based Modeling for Machine Learning

Domain-specific modeling priors and specialized components are becoming increasingly important to the machine learning field. These components integrate specialized knowledge that we have as humans into model. We argue in this thesis that optimization methods provide an expressive set of operations that should be part of the machine learning practitioner's modeling toolbox.

We present two foundational approaches for optimization-based modeling: 1) the OptNet architecture that integrates optimization problems as individual layers in larger end-to-end trainable deep networks, and 2) the input-convex neural network (ICNN) architecture that helps make inference and learning in deep energy-based models and structured prediction more tractable.

We then show how to use the OptNet approach 1) as a way of combining model-free and model-based reinforcement learning with differentiable optimal control and 2) for top-k learning problems. We conclude by showing how to turn the cvxpy domain specific language into a differentiable optimization layer that enables rapid prototyping of the approaches in this thesis.

Thesis Committee:
J. Zico Kolter (Chair)
Barnabas Poczos
Jeff Schneider
Vladlen Koltun (Intel Labs)

For More Information, Contact:


Thesis Oral