Neural Network Compression - The Functional Perspective (+ Extensions)

Israel Mason-Williams presented and discussed the paper Neural Network Compression: The Functional Perspective (+ Extensions) which was accepted at PML4LRS Workshop at ICLR 2024 (Mason-Williams, 2024) and with extensions accepted at NeurIPS Sci4DL Workshop (Mason-Williams et al., 2024).

Abstract

Compression techniques, such as Knowledge distillation, Pruning, and Quantization reduce the computational costs of model inference and enable on-edge machine learning. The efficacy of compression methods is often evaluated through the proxy of accuracy and loss to understand similarity of the compressed model. This study aims to explore the functional divergence between compressed and uncompressed models. The results indicate that Quantization and Pruning create models that are functionally similar to the original model. In contrast, Knowledge distillation creates models that do not functionally approximate their teacher models. The compressed model resembles the dissimilarity of function observed in independently trained models. Therefore, it is verified, via a functional understanding, that Knowledge distillation is not a compression method. Thus, leading to the definition of Knowledge distillation as a training regulariser given that no knowledge is distilled from a teacher to a student.

References

  1. NEURAL NETWORK COMPRESSION: THE FUNCTIONAL PERSPECTIVE
    Israel Mason-Williams
    In 5th Workshop on practical ML for limited/low resource settings, 2024
  2. Knowledge Distillation: The Functional Perspective
    Israel Mason-Williams, Gabryel Mason-Williams, and Mark Sandler
    In NeurIPS 2024 Workshop on Scientific Methods for Understanding Deep Learning, 2024



Enjoy Reading This Article?

Here are some more articles you might like to read next:

  • Agentic AI systems - Promises, Risks and the Paths Forward
  • A Survey of Cognitive Distortion Detection and Classification in NLP
  • Reproducibility The New Frontier in AI Governance
  • The Biology of a Large Language Model (Anthropic)
  • Reflections on The International AI Safety Report