Microsoft Research has introduced BitNet a4.8, a technique that improves the efficiency of 1-bit large language models (LLMs).LLMs use a limited number of bits to represent model weights, reducing memory and computational resources needed.BitNet a4.8 combines hybrid quantization and sparsification techniques to optimize 1-bit LLMs.BitNet a4.8 achieves comparable performance to previous models while using less compute and memory.