Model Quantization - Model Compression for On-Device Deployment - ProjectAI