distinguished lecture series presents

Donald Goldfarb

Columbia University

Research Area

Mathematical optimization and numerical analysis


Tuesday, May 17, 2016 to Thursday, May 19, 2016


MS 6627

Optimization for Learning and Big Data

Many problems in both supervised and unsupervised machine learning (e.g., logistic regression, support vector machines, deep neural networks, robust principal component analysis, dictionary learning, latent variable models) and signal processing (e.g., face recognition and compressed sensing) are solved by optimization and related algorithms. In today’s age of big data, the size of these problems is often formidable. E.g., in logistic regression the objective function may be expressed as the sum of ~10^9 functions (one for each data point) involving ~10^6 variables (features). In this series of talks, we will review current optimization approaches for addressing this challenge from the following classes of methods: first-order (and accelerated variants), stochastic gradient and second-order extensions, alternating direction methods for structured problems (including proximal and conditional gradient and multiplier methods), tensor decomposition, randomized methods for linear systems, and parallel and distributed variants.
Notes and recording available below. Additional recording: https://youtu.be/La0A7Zk8Q4M (video for Lecture 1, Slide 50)
Notes available below.
Notes available below.
recordings & notes
Lecture 1
Lecture 2
Lecture 3
2300 Murphy Hall - Box 951438 - Los Angeles, CA 90095-1438 © 2018