Quantization – Dmytro Dzhulgakov

It’s important to make efficient use of both server-side and on-device compute resources when developing ML applications. To support more efficient deployment on servers and edge devices, PyTorch 1.3 now supports 8-bit model quantization using the familiar eager mode Python API.

Source of this PyTorch AI Video

AI video(s) you might be interested in …