A Geometrical Connection Between Sparse and Low-rank Matrices

Computer Science Distinguished Lecture Series
Monday, October 24, 2022
11:40 AM - 1:00 PM
Department of Computer Science, 500 W. 120th St., New York, New York 10027
Room/Area: CSB 451 Computer Science Auditorium
Add to Calendar

Link added to clipboard:

https://events.columbia.edu/cal/event/eventView.do?b=de&calPath=%2Fpublic%2Fcals%2FMainCal&guid=CAL-00bbdb71-83d06435-0183-d2336d69-00003168events@columbia.edu&recurrenceId=

Computer Science Distinguished Lecture Series

Lawrence Saul, Flatiron Institute

A Geometrical Connection Between Sparse and Low-rank Matrices and Its Uses for Machine Learning

Abstract: Many problems in high dimensional data analysis can be formulated as a search for structure in large matrices. One important type of structure is sparsity; for example, when a matrix is sparse, with a large number of zero elements, it can be stored in a highly compressed format. Another type of structure is linear dependence; when a matrix is low-rank, it can be expressed as the product of two smaller matrices. It is well known that neither one of these structures implies the other. But can one find more subtle connections by looking beyond the canonical decompositions of linear algebra?

In this talk, I will consider when a sparse nonnegative matrix can be recovered from a real-valued matrix of significantly lower rank. Of particular interest is the setting where the positive elements of the sparse matrix encode the similarities of nearby points on a low-dimensional manifold. The recovery can then be posed as a problem in manifold learning—namely, how to learn a similarity-preserving mapping of high-dimensional inputs into a lower-dimensional space. I will describe an algorithm for this problem based on a generalized low-rank decomposition of sparse matrices. This decomposition has the interesting property that it can be encoded by a neural network with one layer of rectified linear units; since the algorithm discovers this encoding, it can also be viewed as a layerwise primitive for deep learning. Finally, I will apply the algorithm to data sets where vector magnitudes and small cosine distances have interpretable meanings (e.g., the brightness of an image, the similarity to other words). On these data sets, the algorithm is able to discover much lower dimensional representations that preserve these meanings.

Bio: Lawrence Saul is a Senior Research Scientist in the Center for Computational Mathematics (CCM) at the Flatiron Institute. He joined CCM in July 2022 as a group leader in machine learning; previously, he was a Professor and Vice Chair in the Department of Computer Science and Engineering at UC San Diego.

Event Contact Information:
Daniel Hsu
LOCATION:
  • Morningside
TYPE:
  • Lecture
CATEGORY:
  • Computer Science
EVENTS OPEN TO:
  • Alumni
  • Faculty
  • Family-friendly
  • Graduate Students
  • Postdocs
  • Prospective Students
  • Public
  • Staff
  • Students
  • Trainees
BACK TO EVENTS

Date Navigation Widget

Filter By

Subscribe Export Options

Getting to Columbia

Other Calendars

Guests With Disabilities