Gaussian distributed weights for LLMs
This post examines the NF4 4-bit floating point format and higher precision analogs used for quantizing LLM weights. NF4 and FP4 are common 4-bit data types in bitsandbytes, often found in weights downloaded from Hugging Face.