Joint IAS/PNI Seminar on ML and Neuroscience

A geometric data structure from neuroscience

An intriguing primitive representation, "expand-and-sparsify", appears in the olfactory system of the fly and the sensory systems of several other organisms. It maps an input vector to a much higher-dimensional sparse representation, using a random linear transformation followed by winner-take-all thresholding. I'll show that this representation has a variety of formal properties, such as (1) locality preservation, (2) universal approximation, and (3) adaptivity to manifold structure, that make it an attractive data structure for algorithms and machine learning. In particular, mimicking the fly's circuitry yields algorithms for similarity search and for novelty detection that have provable guarantees as well as being effective in practice. This talk is based on work with Saket Navlakha (Salk Institute), Chuck Stevens (Salk Institute), and Chris Tosh (Columbia). The published results appear in these two papers: S. Dasgupta, C.F. Stevens and S. Navlakha. A neural algorithm for a fundamental computing problem. Science 2017. https://science.sciencemag.org/content/sci/358/6364/793 S. Dasgupta, T.C. Sheehan, C.F. Stevens and S. Navlakha. A neural data structure for novelty detection. PNAS, 2018. https://www.pnas.org/content/115/51/13093

Date & Time

October 01, 2019 | 12:00pm – 1:30pm

Location

White-Levy

Affiliation

University of California, San Diego; Special Year Member, School of Mathematics

Categories