AI Seminar:  A Randomized and Provably Easy Construction of High-Accuracy Binary/Quantized Neural Networks

Alireza Aghasi
Event Speaker
Alireza Aghasi
Associate Professor, School of Electrical Engineering and Computer Science, ¾«¶«Ó°ÊÓ State University
Event Type
Artificial Intelligence
Date
Event Location
Zoom and LINC 302
Event Description

Training binary or quantized neural networks is inherently challenging due to the discrete nature of the problem and the high dimensionality of modern models. But imagine the existence of a meta-distribution from which one could easily sample binary neural networks with controllable accuracy—and the sampling process did not involve any training in the binary/quantized domain. We present a novel randomized framework that makes this possible. In contrast to traditional quantization and hyperdimensional computing methods, our approach models binary representations as points on the Hamming cube, embedding data directly into binary space. At the core of our method is Grothendieck’s lemma, which enables the construction of a new class of network layers—Grothendieck layers—that emulate standard neural layers in the floating-point domain. Each of these layers admits a randomized binary counterpart—an embedded hyperdimensional network—whose accuracy closely mirrors that of the original, with rigorous theoretical guarantees derived from the concentration of measure in high dimensions. Empirically, these sampled binary networks match the performance of convolutional neural networks while being amenable to efficient deployment on edge devices. This framework introduces a principled and scalable pathway for designing robust binary networks without requiring any discrete optimization during training.
 

Speaker Biography

Alireza Aghasi joined the School of Electrical Engineering and Computer Science at ¾«¶«Ó°ÊÓin Fall 2022. Between 2017 and 2022, he was an assistant professor in the Department of Data Science and Analytics at the Robinson College of Business, Georgia State University. Prior to this position he was a research scientist with the Department of Mathematical Sciences, IBM T.J. Watson research center, Yorktown Heights. From 2015 to 2016 he was a postdoctoral associate with the computational imaging group at the Massachusetts Institute of Technology, and between 2012 and 2015 he served as a postdoctoral research scientist with the compressed sensing group at Georgia Tech. His research fundamentally focuses on optimization theory and statistics, with applications to various areas of data science, artificial intelligence, modern signal processing and physics-based inverse problems.