SEEM 5380: Optimization Methods for High-Dimensional Statistics
2016-17 Second Term
Announcements
General Information
Course Description
The prevalence of high-dimensional data has motivated active research on efficient methods for tackling optimization problems that arise in statistical analysis. In this course, we will give an introduction to this exciting area of research, with emphasis on the theory of structured regularizers for high-dimensional statistics and the design and analysis of statistically and computationally efficient optimization algorithms. Applications in various areas of science and engineering, such as machine learning, signal processing, and statistics, will also be discussed. Prerequisite: ENGG 5501 or equivalent.
Course Outline (subject to change)
Course Requirements
Homework sets (60%) and a take-home final examination (40%)
Open problems will be introduced throughout the semester. Students who solve any one of the open problems will automatically receive an A for the course.
General References
Schedule and Reading
Jan 17: Bound on statistical error for general models. Application to sparse linear regression.
Reading:
Reading:
Foucart, Rauhut. A Mathematical Introduction to Compressive Sensing. Springer Science+Business Media, New York, 2013.
Rudelson, Zhou. Reconstruction from Anisotropic Random Measurements. IEEE Transactions on Information Theory 59(6): 3434-3447, 2013.
Zhou. Restricted Eigenvalue Conditions on Subgaussian Random Matrices. Technical Report, 2009.
Negahban, Wainwright. Restricted Strong Convexity and Weighted Matrix Completion: Optimal Bounds with Noise. Journal of Machine Learning Research 13: 1665-1697, 2012.
Reading:
Hou, Zhou, So, Luo. On the Linear Convergence of the Proximal Gradient Method for Trace Norm Regularization. Advances in Neural Information Processing Systems 26: Proceedings of the 2013 Conference (NIPS 2013), pp. 710-718, 2013.
See, in particular, Theorem 3.2.
WARNING: The conclusion of Theorem 3.1 of the above paper is not correct. This is due to a flaw in the proof. The correct conclusion appears in Proposition 12 of
Zhou, So. A Unified Approach to Error Bounds for Structured Convex Optimization Problems. Mathematical Programming, Series A, to appear, 2016.
Combettes, Wajs. Signal Recovery by Proximal Forward-Backward Splitting. Multiscale Modeling and Simulation 4(4): 1168-1200, 2005.
Rockafellar, Wets. Variational Analysis. Grundlehren der Mathematischen Wissenschaften 327, Springer-Verlag, 2004. (Particularly Chapters 1.G and 2.E)
The lemma on p.6 of the notes (concerning certain monotonicity properties of the proximal map) first appears in
Sra. Scalable Nonconvex Inexact Proximal Splitting. Advances in Neural Information Processing Systems 25: Proceedings of the 2012 Conference (NIPS 2012), pp. 530-538, 2012.
However, to obtain the desired conclusion, an envelope theorem should be invoked in the proof. In fact, just having a unique solution to displayed equation (18) of the above paper is not sufficient to guarantee the differentiability of the function Eg. The envelope theorem mentioned on p.8 of the notes comes from Proposition 1 of
Oyama, Takenawa. On the (Non-)Differentiability of the Optimal Value Function When the Optimal Solution Is Unique. Manuscript, 2013.
Envelope theorems have many applications in economics; see the above paper and the pointers therein for details.
Agarwal, Negahban, Wainwright. Fast Global Convergence of Gradient Methods for High-Dimensional Statistical Recovery. Annals of Statistics 40(5): 2452-2482, 2012. Supplementary Material
The above paper shows that for several classes of regularized loss minimization problems, certain first-order methods enjoy global linear convergence up to the statistical precision of the model. Modulo the global vs. local nature of the linear convergence, this is weaker than the convergence result derived from the error bound theory. Nevertheless, the approach developed in the above paper can be used to tackle problems that are not known to possess an error bound.
Reading:
Zhou, So. A Unified Approach to Error Bounds for Structured Convex Optimization Problems. Mathematical Programming, Series A, to appear, 2016.
Güler. Foundations of Optimization. Graduate Texts in Mathematics 258, Springer Science+Business Media, LLC, 2010.
Bauschke, Borwein. On Projection Algorithms for Solving Convex Feasibility Problems. SIAM Review 38(3): 367-426, 1996.
Dontchev, Rockafellar. Implicit Functions and Solution Mappings. Springer Monographs in Mathematics, Springer Science+Business Media, LLC, 2009.
Reading:
Yue, Zhou, So. A Family of SQA Methods for Non-Smooth Non-Strongly Convex Minimization with Provable Convergence Guarantees. Manuscript, 2017.
Reading:
Pilanci, Wainwright. Newton Sketch: A Near Linear-Time Optimization Algorithm with Linear-Quadratic Convergence. SIAM Journal on Optimization, to appear, 2016.
As noted in the lecture, the approach in the above paper does not offer any advantage for unconstrained problems, as the sketching dimension could be as high as the dimension of the decision vector. In fact, the same is true for a constrained problem if the optimal solution lies in the interior of the constraint set; see displayed equation (3.8) in the above paper.
Reading:
Loh, Wainwright. High-Dimensional Regression with Noisy and Missing Data: Provable Guarantees with Nonconvexity. Annals of Statistics 40(3): 1637-1664, 2012. Supplementary Material
Mar 14: Bound on statistical error in the non-convex setting.
Reading:
Fan, Li. Variable Selection via Nonconcave Penalized Likelihood and Its Oracle Properties. Journal of the American Statistical Association 96(456): 1348-1360, 2001.
Zhang. Nearly Unbiased Variable Selection under Minimax Concave Penalty. Annals of Statistics 38(2): 894-942, 2010.
For an in-depth analysis of the role of various non-convex regularizers in sparse estimation problems, see
Zhang, Zhang. A General Theory of Concave Regularization for High-Dimensional Sparse Estimation Problems. Statistical Science 27(4): 576-593, 2012. Supplementary Material
Loh, Wainwright. Regularized M-estimators with Nonconvexity: Statistical and Algorithmic Theory for Local Optima. Journal of Machine Learning Research 16(Mar): 559-616, 2015.
Reading:
Loh, Wainwright. Regularized M-estimators with Nonconvexity: Statistical and Algorithmic Theory for Local Optima. Journal of Machine Learning Research 16(Mar): 559-616, 2015.
Loh, Wainwright. High-Dimensional Regression with Noisy and Missing Data: Provable Guarantees with Nonconvexity. Annals of Statistics 40(3): 1637-1664, 2012. Supplementary Material
Mar 28: The Generalized Power Method and bound on estimation error.
Reading:
Liu, Yue, So. On the Estimation Performance and Convergence Rate of the Generalized Power Method for Phase Synchronization. Preprint, 2016.
Absil, Mahony, Sepulchre. Optimization Algorithms on Matrix Manifolds. Princeton University Press, 2008.
Apr 4: Ching Ming Festival.
Reading:
Zhong, Boumal. Near-Optimal Bounds for Phase Synchronization. arXiv, 2017.
Apr 11: The Generalized Power Method for MIMO Detection.
Reading:
So. Probabilistic Analysis of the Semidefinite Relaxation Detector in Digital Communications. Proceedings of the 21st Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2010), pp. 698-711, 2010.
Apr 18: Wrap-up and possible future studies.
Reading:
So, Zhou. Non-Asymptotic Convergence Analysis of Inexact Gradient Methods for Machine Learning Without Strong Convexity. Accepted for publication in Optimization Methods and Software, 2017.
Sun, Qu, Wright. When are Nonconvex Problems not Scary? Preprint, 2016.
Ge, Jin, Zheng. No Spurious Local Minima in Nonconvex Low Rank Problems: A Unified Geometric Analysis. Manuscript, 2017.
Yin, Lou, He, Xin. Minimization of l1-2 for Compressed Sensing. SIAM Journal on Scientific Computing 37(1): A536-A563, 2015.
Pang, Razaviyayn, Alvarado. Computing B-Stationary Points of Nonsmooth DC Programs. Mathematics of Operations Research 42(1): 95-118, 2017.
Nouiehed, Pang, Razaviyayn. On the Pervasiveness of Difference-Convexity in Optimization and Statistics Manuscript, 2017.
Homework Sets
Last Updated: April 27, 2017