1 hour ago · 0 comments

The previous post looked at the FP4 4-bit floating point format. This post will look at another 4-bit floating point format, NF4, and higher precision analogs. NF4 and FP4 are common bitsandbytes 4-bit data types. If you download LLM weights from Hugging Face quantized to four bits, the weights might be in NF4 or FP4 […] The post Gaussian distributed weights for LLMs first appeared on John D. Cook.

No comments yet. Log in to reply on the Fediverse. Comments will appear here.