Neural Network Quantization and Compression with Tijmen Blankevoort - TWIML Talk #292
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) - A podcast by Sam Charrington - Mondays
![](https://is5-ssl.mzstatic.com/image/thumb/Podcasts116/v4/0c/c1/26/0cc12665-42ab-fe43-5bf4-26707e3b2fcb/mza_11206870036036003114.jpeg/300x300bb-75.jpg)
Categories:
Today we’re joined by Tijmen Blankevoort, a staff engineer at Qualcomm, who leads their compression and quantization research teams. In our conversation with Tijmen we discuss: • The ins and outs of compression and quantization of ML models, specifically NNs, • How much models can actually be compressed, and the best way to achieve compression, • We also look at a few recent papers including “Lottery Hypothesis."