Site menu:

picture by

OPT2021

We welcome you to participate in the 13th International (Virtual) OPT Workshop on Optimization for Machine Learning, to be held as a part of the NeurIPS 2021 conference. This year we particularly encourage (but not limit) submissions in the area of Beyond Worst-case Complexity.


We are looking forward to an exciting OPT 2021!


All events online

Schedule

Time Speaker Title
6:15am-6:55amWelcome social event in gather.town (lounge). Login and say hello!


Session 1 (Moderator: Sebastian Stich, co-moderator: Martin Takáč)

6:55am-7:00amOrganizers Opening Remarks to Session 1
7:00am-7:30amShai Shalev-Shwartz (Hebrew University of Jerusalem)
(video stream)
Deep Learning: Success, Failure, and the Border between them[abstract]
7:30am-8:00amMartin Jaggi (EPFL)
(video stream)
Learning with Strange Gradients[abstract]
8:00am-8:30amContributed talks
(live stream)
Oral (10min):
  • Futong Liu (EPFL): Understanding Memorization from the Perspective of Optimization via Efficient Influence Estimation
Spotlights (5min):
  • Abdurakhmon Sadiev (MIPT): Decentralized Personalized Federated Learning: Lower Bounds and Optimal Algorithm for All Personalization Modes
  • Frederik Benzing (ETH Zurich): Fast, Exact Subsampled Natural Gradients and First-Order KFAC
  • Simon Roburin (valeo.ai / imagine ENPC): Spherical Perspective on Learning with Normalization Layers
[papers]
8:30am-9:55amPoster Session 1 and Break Join for the poster session on gather.town (room 1).
Authors will present their work.
[posters]


Session 2 (Moderator: Katya Scheinberg, co-moderator: Courtney Paquette)

9:55am-10:00amOrganizers Opening Remarks
10:00am-10:30amCoralia Cartis (University of Oxford)
(video stream)
The global optimization of functions with low effective dimension - better than worst-case?[abstract]
10:30am-11:00amCristóbal Guzmán (University of Twente)
(video stream)
Non-Euclidean Differentially Private Stochastic Convex Optimization[abstract]
11:00am-11:30amContributed talks
(live stream)
Oral (10min):
  • Junchi Li (UC Berkeley): On the convergence of stochastic extragradient for bilinear games using restarted iteration averaging
Spotlights (5min):
  • Jeffery Kline (American Family Insurance): Farkas' Theorem of the Alternative for Prior Knowledge in Deep Networks
  • Lyle Kim (Rice University): Fast, Acceleration and Stability of the Stochastic Proximal Point Algorithm
  • Pascal Esser (Technical University of Munich): Towards Modeling and Resolving Singular Parameter Spaces using Stratifolds
[papers]
11:30am-12:55pmBreak Join the gather.town lounge, or the poster rooms, (room 1) (room 2) for discussions with other participants.
[posters]


Session 3 (Moderator: Oliver Hinder, co-moderator: Courtney Paquette)

12:55pm-1:00pmOrganizers Opening Remarks
1:00pm-1:30pmDamek Davis (Cornell University)
(video stream)
Avoiding saddle points in nonsmooth optimization[abstract]
1:30pm-2:00pmJelena Diakonikolas (University of Wisconsin-Madison)
(video stream)
Faster Empirical Risk Minimization[abstract]
2:00pm-2:30pmContributed talks
(live stream)
Oral (10min):
  • Wenhao Zhan (Princeton University): Policy Mirror Descent for Regularized RL: A Generalized Framework with Linear Convergence
Spotlights (5min):
  • Akhilesh Soni (University of Wisconsin-Madison): Integer Programming Approaches To Subspace Clustering With Missing Data
  • Boyue Li (Carnegie Mellon University): DESTRESS: Computation-Optimal and Communication-Efficient Decentralized Nonconvex Finite-Sum Optimization
  • Grigory Malinovsky (KAUST): Better Linear Rates for SGD with Data Shuffling
[papers]
2:30pm-15:55pmPoster Session 2 and Break Join for the poster session on gather.town (room 2).
Authors will present their work.
[posters]


Session 4 (Moderator: Quanquan Gu, co-moderator: Oliver Hinder)

3:55pm-4:00pmOrganizers Opening Remarks
4:00pm-4:30pmYinyu Ye (Stanford University)
(video stream)
Online Learning via Linear Programming[abstract]
4:30pm-5:00pmMichael Mahoney (UC Berkeley)
(video stream)
Putting Randomized Matrix Algorithms in LAPACK, and Connections with Second-order Stochastic Optimization[abstract]
5:00pm-5:30pmContributed talks
(live stream)
Oral (10min):
  • Agnieszka Słowik (University of Cambridge): On the Relation between Distributionally Robust Optimization and Data Curation
Spotlights (5min):
  • Jacques Chen (University of British Columbia): Heavy-tailed noise does not explain the gap between SGD and Adam on Transformers
  • Neha Wadia (UC Berkeley): Optimization with Adaptive Step Size Selection from a Dynamical Systems Perspective
  • Difan Zou (UC Los Angeles): Understanding the Generalization of Adam in Learning Neural Networks with Proper Regularization
[papers]


Closing Remarks

5:30pm-5:35pmOrganizers Closing Remarks