Deep Learning Toolbox Model Quantization Library
Quantize and compress deep learning models
2,2K Downloads
Aktualisiert
16. Okt 2024
Deep Learning Toolbox Model Quantization Library enables quantization and compression of your deep learning models to reduce the memory footprint and computational requirements of your deep neural network.
Quantization to INT8 is supported for CPUs, FPGAs, and NVIDIA GPUs, for supported layers. The library enables you to collect layer level data on the weights, activations, and intermediate computations. Using this data, the library quantizes your model and provides metrics to validate the accuracy of the quantized network against the single precision baseline. The iterative workflow allows you to optimize the quantization strategy.
The library also supports structural compression of models with pruning and projection. Both techniques reduce the sizes of deep neural networks by removing elements that have the smallest impact on inference accuracy.
Please refer to the documentation here: https://www.mathworks.com/help/deeplearning/quantization.html
Quantization Workflow Prerequisites can be found here:
If you have download or installation problems, please contact Technical Support - www.mathworks.com/contact_ts
Additional Resources
- Learn more about MATLAB and Simulink for TinyML
- Quantization Aware Training (QAT) with MobileNet-v2 (Example, GitHub Repo)
- Overview Video - https://www.youtube.com/watch?v=jufOpBeSvHM
Kompatibilität der MATLAB-Version
Erstellt mit
R2020a
Kompatibel mit R2020a bis R2024b
Plattform-Kompatibilität
Windows macOS (Apple Silicon) macOS (Intel) LinuxKategorien
Mehr zu Deep Learning Toolbox finden Sie in Help Center und MATLAB Answers
Tags
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!Live Editor erkunden
Erstellen Sie Skripte mit Code, Ausgabe und formatiertem Text in einem einzigen ausführbaren Dokument.