Quantcast
Channel: How the "quantizantion" of models does work? - Artificial Intelligence Stack Exchange
Viewing all articles
Browse latest Browse all 2

How the "quantizantion" of models does work?

$
0
0

Consider that a model using FP16 precision is quantized to a lower precision like INT8. Does this reduce the accuracy of the model? From what I know it is designed to reduce the size and required RAM to run the models.


Viewing all articles
Browse latest Browse all 2

Latest Images

Trending Articles



Latest Images

<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>
<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596344.js" async> </script>