Maximum output discrepancy computation for convolutional neural network compression

Zihao Mo, Weiming Xiang

Research output: Contribution to journalArticlepeer-review

Abstract

Network compression methods minimize the number of network parameters and computation costs while maintaining desired network performance. However, the safety assurance of many compression methods is based on a large amount of experimental data, whereas unforeseen incidents beyond the experiment data may result in unsafe consequences. In this work, we developed a discrepancy computation method for two convolutional neural networks by giving a concrete value to characterize the maximum output difference between the two networks after compression. Using Imagestar-based reachability analysis, we propose a novel method to merge the two networks to compute the difference. We illustrate reachability computation for each layer in the merged network, such as the convolution, max pooling, fully connected, and ReLU layers. We apply our method to a numerical example to prove its correctness. Furthermore, we implement our developed methods on the VGG16 model with the Quantization Aware Training (QAT) compression method; the results show that our approach can efficiently compute the accurate maximum output discrepancy between the original neural network and the compressed neural network.

Original languageEnglish (US)
Article number120367
JournalInformation Sciences
Volume665
DOIs
StatePublished - Apr 2024

Keywords

  • Convolutional neural network
  • Discrepancy computation
  • Neural network compression
  • Reachability analysis

ASJC Scopus subject areas

  • Software
  • Control and Systems Engineering
  • Theoretical Computer Science
  • Computer Science Applications
  • Information Systems and Management
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Maximum output discrepancy computation for convolutional neural network compression'. Together they form a unique fingerprint.

Cite this