Optimization Techniques - W2023 - Lecture 10 (Distributed Optimization & Non-Smooth Optimization)
The course "Optimization Techniques" (ENGG*6140, section 2) at the School of Engineering at the University of Guelph. Instructor: Benyamin Ghojogh Lecture 10 continues the distributed optimization methods by reviewing ADMM introduced in the previous session. Then, we talk about how to make univariate/multivariate optimization problem distributed. An example of ADMM in image processing is also introduced. Then, we switch to non-smooth optimization. We start with lasso (l1 norm) optimization. Then, we review the various methods for solving non-smooth optimization. These methods include convex conjugate, Huber and pseudo-Huber functions, soft-thresholding and proximal methods, coordinate descent, and subgradient methods. The slides are available at: https://bghojogh.github.io/pages/uoguelph/engg-6140-w23/ Very useful optimization courses at Stanford University by Prof. Stephen Boyd: https://www.youtube.com/playlist?list=PL3940DD956CDF0622 Chapters: 0:00 - Talking about project presentation 4:30 - Correcting a typo in Wolfe conditions for line-search 7:13 - Brief review of distributed optimization from previous session 12:33 - Use cases of distributed algorithms 14:34 - Making univariate optimization problem distributed 20:12 - Making multivariate optimization problem distributed 32:42 - Example of ADMM in image processing 44:22 - Important references in distributed optimization 48:00 - Non-smooth optimization (introduction) 50:27 - Lasso (l1 norm) regularization 1:01:33 - Convex conjugate 1:09:26 - Convex conjugate of l1 norm 1:15:01 - Gradient in terms of convex conjugate 1:17:33 - Proximity function and Huber function 1:25:23 - Huber and pseudo-Huber functions 1:33:21 - Soft-thresholding and proximal methods 1:38:31 - Coordinate descent 1:47:19 - Coordinate descent for l1 norm optimization 1:58:14 - Subgradient 2:04:15 - Subdifferential 2:06:47 - Subdifferential for l1 norm 2:11:33 - First-order optimality condition for non-smooth function 2:15:09 - Subgradient in example functions 2:18:23 - Subgradient method 2:22:00 - Stochastic subgradient method 2:23:23 - Projected subgradient method 2:24:32 - Important references in non-smooth optimization 2:29:13 - Talking about the next sessions
Download
0 formatsNo download links available.