Department of Mathematics

Applied Mathematics

  •  Rayan Saab, University of California, San Diego (UCSD)
  •  Quantizing neural networks
  •  09/22/2022
  •  2:30 PM - 3:30 PM
  •  Online (virtual meeting) (Virtual Meeting Link)
  •  Mark A Iwen (iwenmark@msu.edu)

Neural networks are highly non-linear functions often parametrized by a staggering number of weights. Miniaturizing these networks and implementing them in hardware is a direction of research that is fueled by a practical need, and at the same time connects to interesting mathematical problems. For example, by quantizing, or replacing the weights of a neural network with quantized (e.g., binary) counterparts, massive savings in cost, computation time, memory, and power consumption can be attained. Of course, one wishes to attain these savings while preserving the action of the function on domains of interest. We present data-driven and computationally efficient methods for quantizing the weights of already trained neural networks and we prove that our methods have favorable error guarantees under a variety of assumptions. We also discuss extensions and provide the results of numerical experiments, on large multi-layer networks, to illustrate the performance of our methods. Time permitting, we will also discuss open problems and related areas of research.

 

Contact

Department of Mathematics
Michigan State University
619 Red Cedar Road
C212 Wells Hall
East Lansing, MI 48824

Phone: (517) 353-0844
Fax: (517) 432-1562

College of Natural Science