Geometric Optimization in Machine Learning

Suvrit Sra, Reshad Hosseini

Research output: Chapter in Book/Report/Conference proceedingChapterpeer-review

9 Scopus citations

Abstract

Machine learning models often rely on sparsity, low-rank, orthogonality, correlation, or graphical structure. The structure of interest in this chapter is geometric, specifically the manifold of positive definite (PD) matrices. Though these matrices recur throughout the applied sciences, our focus is on more recent developments in machine learning and optimization. In particular, we study (i) models that might be nonconvex in the Euclidean sense but are convex along the PD manifold; and (ii) ones that are neither Euclidean nor geodesic convex but are nevertheless amenable to global optimization. We cover basic theory for (i) and (ii); subsequently, we present a scalable Riemannian limited-memory BFGS algorithm (that also applies to other manifolds). We highlight some applications from statistics and machine learning that benefit from the geometric structure studies.

Original languageEnglish
Title of host publicationAdvances in Computer Vision and Pattern Recognition
PublisherSpringer Science and Business Media Deutschland GmbH
Pages73-91
Number of pages19
DOIs
StatePublished - 2016
Externally publishedYes

Publication series

NameAdvances in Computer Vision and Pattern Recognition
ISSN (Print)2191-6586
ISSN (Electronic)2191-6594

Fingerprint

Dive into the research topics of 'Geometric Optimization in Machine Learning'. Together they form a unique fingerprint.

Cite this