Deep neural networks have delivered remarkable performance and have been...
Extensive studies have shown that deep learning models are vulnerable to...
Quantization of transformer language models faces significant challenges...
Adversarial training has been demonstrated to be one of the most effecti...
Transformer architecture has become the fundamental element of the wides...
Recently, post-training quantization (PTQ) has driven much attention to
...
Model quantization has emerged as an indispensable technique to accelera...
Model binarization is an effective method of compressing neural networks...
Deep neural networks (DNNs) are vulnerable to adversarial noises, which
...
Systematic error, which is not determined by chance, often refers to the...
Spiking Neural Network (SNN) has been recognized as one of the next
gene...
Quantization has emerged as one of the most prevalent approaches to comp...
We study the challenging task of neural network quantization without
end...
User data confidentiality protection is becoming a rising challenge in t...
The binary neural network, largely saving the storage and computation, s...
Network quantization has rapidly become one of the most widely used meth...
Recently low-bit (e.g., 8-bit) network quantization has been extensively...
Binary neural networks have attracted numerous attention in recent years...
Weight and activation binarization is an effective approach to deep neur...
Hardware-friendly network quantization (e.g., binary/uniform quantizatio...