Gaussian distributed weights for LLMs
The Endeavour 2026-04-18
Summary:
The previous post looked at the FP4 4-bit floating point format. This post will look at another 4-bit floating point format, NF4, and higher precision analogs. NF4 and FP4 are common bitsandbytes 4-bit data types. If you download LLM weights from Hugging Face quantized to four bits, the weights might be in NF4 or FP4 […]
The post Gaussian distributed weights for LLMs first appeared on John D. Cook.