< Back to Article
SensiMix: Sensitivity-Aware 8-bit index & 1-bit value mixed precision quantization for BERT compression
Table 5
Comparison of the sensitivity of Self-Attention layer and FFN in BERT.
The result indicates that Self-Attention (SA) layer is more sensitive than FFN.
doi: https://doi.org/10.1371/journal.pone.0265621.t005