Existing works on Binary Neural Network (BNN) have focused on model weights and activations, neglecting considerations on input raw data.
This article introduces the Generic Learned Thermometer (GLT) encoding technique to enhance input data representation for BNN by learning non-linear quantization thresholds.
GLT involves multiple data binarizations to replace conventional Analog to Digital Conversion (ADC) using natural binary coding, improving model efficiency.
A compact network topology with lightweight grouped convolutions trained through block pruning and Knowledge Distillation is proposed, resulting in small, fully-binarized models suitable for efficient inference tasks.