Joint IAS/PNI Seminar on ML and Neuroscience
A geometric data structure from neuroscience
An intriguing primitive representation, "expand-and-sparsify", appears in the olfactory system of the fly and the sensory systems of several other organisms. It maps an input vector to a much higher-dimensional sparse representation, using a random linear transformation followed by winner-take-all thresholding. I'll show that this representation has a variety of formal properties, such as (1) locality preservation, (2) universal approximation, and (3) adaptivity to manifold structure, that make it an attractive data structure for algorithms and machine learning. In particular, mimicking the fly's circuitry yields algorithms for similarity search and for novelty detection that have provable guarantees as well as being effective in practice. This talk is based on work with Saket Navlakha (Salk Institute), Chuck Stevens (Salk Institute), and Chris Tosh (Columbia). The published results appear in these two papers: S. Dasgupta, C.F. Stevens and S. Navlakha. A neural algorithm for a fundamental computing problem. Science 2017. https://science.sciencemag.org/content/sci/358/6364/793 S. Dasgupta, T.C. Sheehan, C.F. Stevens and S. Navlakha. A neural data structure for novelty detection. PNAS, 2018. https://www.pnas.org/content/115/51/13093