zlacker

[return to "Releasing weights for FLUX.1 Krea"]
1. bangal+y91[view] [source] 2025-07-31 20:52:12
>>vmatsi+(OP)
Can someone ELI5 why the safetensor file is 23.8 GB, given the 12B parameter model? Does the model use closer to 24 GB of VRAM or 12 GB of VRAM. I've always associated a 1 billion parameter = 1 GB of VRAM. Is this estimate inaccurate?
◧◩
2. pipers+aa1[view] [source] 2025-07-31 20:55:08
>>bangal+y91
A parameter can be any size float. Lots of downloadable models are FP8 (8 bits per parameter), but it appears this model is FP16 (16 bits per parameter)

Often, the training is done in FP16 then quantized down to FP8 or FP4 for distribution.

◧◩◪
3. dragon+eG1[view] [source] 2025-08-01 01:10:23
>>pipers+aa1
I think they are bfloat16, not FP16, but they are both 16bpw formats, so it doesn't make a size difference.
◧◩◪◨
4. iyn+Dh2[view] [source] 2025-08-01 08:38:44
>>dragon+eG1
Wiki article on bfloat16 for reference, since it was new to me: https://en.wikipedia.org/wiki/Bfloat16_floating-point_format
[go to top]