Back to course
Back to home
Interactive Quiz
Test your knowledge!
1
What is the main difference between symmetric quantization and asymmetric quantization?
A
Symmetric quantization uses an offset zero-point while asymmetric quantization maps zero to zero.
B
Symmetric quantization maps the floating zero to the quantized zero, whereas asymmetric quantization maps the floating minimum and maximum range to the quantized range with an offset zero-point.
C
Asymmetric quantization is always faster than symmetric quantization.
D
Symmetric quantization is used only for activations, asymmetric quantization only for weights.
2
What is the main advantage of Quantization Aware Training (QAT) compared to Post-Training Quantization (PTQ)?
A
QAT allows quantization without precision loss even at 1 bit.
B
QAT integrates the quantization process during training, allowing the model to adapt to quantization-induced errors.
C
PTQ requires fewer computations than QAT, making QAT unnecessary for large models.
D
QAT is only compatible with symmetric quantization.
3
What is the main purpose of calibration in the quantization process?
A
Adjust the model weights to optimize accuracy.
B
Find the range of values for activations or weights that minimizes quantization error.
C
Reduce the model size by decreasing the number of bits used.
D
Modify the model structure to improve inference speed.
4
What is the primary reason for using clipping before quantization?
A
To increase the precision of quantized weights by reducing the dynamic range.
B
To reduce the effect of extreme values (outliers) that distort the mapping and increase quantization error.
C
To ensure that quantization is always symmetric around zero.
D
To speed up the calculation of the scale factor and zero-point.
5
What is the main difference between dynamic quantization and static quantization of activations?
A
Dynamic quantization calculates the scale factor and zero-point before inference, while static quantization does so during inference.
B
Static quantization requires a calibration dataset to compute the scale factor and zero-point before inference, whereas dynamic quantization calculates them during inference for each layer.
C
Dynamic quantization is always faster than static quantization.
D
Static quantization can only be used for weights, not for activations.
Score: 0/5
Score: 0/5