Guaranteed Quantization Error Computation for Neural Network Model Compression

Wesley Cooke, Zihao Mo, Weiming Xiang

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Scopus citations

Abstract

Neural network model compression techniques can address the computation issue of deep neural networks on embedded devices in industrial systems. The guaranteed output error computation problem for neural network compression with quantization is addressed in this paper. A merged neural network is built from a feedforward neural network and its quantized version to produce the exact output difference between two neural networks. Then, optimization-based methods and reachability analysis methods are applied to the merged neural network to compute the guaranteed quantization error. Finally, a numerical example is proposed to validate the applicability and effectiveness of the proposed approach.

Original languageEnglish (US)
Title of host publication2023 IEEE International Conference on Industrial Technology, ICIT 2023
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350336504
DOIs
StatePublished - 2023
Event2023 IEEE International Conference on Industrial Technology, ICIT 2023 - Orlando, United States
Duration: Apr 4 2023Apr 6 2023

Publication series

NameProceedings of the IEEE International Conference on Industrial Technology
Volume2023-April

Conference

Conference2023 IEEE International Conference on Industrial Technology, ICIT 2023
Country/TerritoryUnited States
CityOrlando
Period4/4/234/6/23

Keywords

  • model compression
  • neural networks
  • quantization

ASJC Scopus subject areas

  • Computer Science Applications
  • Electrical and Electronic Engineering

Fingerprint

Dive into the research topics of 'Guaranteed Quantization Error Computation for Neural Network Model Compression'. Together they form a unique fingerprint.

Cite this