Machine Learning/Edge Device

ONNX Quantization Test

light_meal 2023. 3. 18. 11:36
728x90

mobileNet float32 → uInt8

  • 예제

https://github.com/microsoft/onnxruntime-inference-examples

 

GitHub - microsoft/onnxruntime-inference-examples: Examples for using ONNX Runtime for machine learning inferencing.

Examples for using ONNX Runtime for machine learning inferencing. - GitHub - microsoft/onnxruntime-inference-examples: Examples for using ONNX Runtime for machine learning inferencing.

github.com

quantization → notebook → imagenet_v2 → mobilenet.ipynb

 

  • memory size
# ONNX full precision model size (MB)
13.31911563873291

# ONNX quantized model size (MB)
3.4079103469848633

 

  • runtime
# ONNX full precision model
tabby 0.70594245
Egyptian cat 0.14357993
tiger cat 0.12944907
lynx 0.0051510707
plastic bag 0.0037339667

# ONNX quantized model size
tabby 0.65022784
tiger cat 0.1949021
Egyptian cat 0.14421292
lynx 0.0047474923
tiger 0.0021263342
728x90