Comment by mchiang 8 days ago OpenAI has only provided MXFP4 weights. These are the same weights used by other cloud providers. 4 comments mchiang Reply irthomasthomas 8 days ago Oh, I didn't know that. Weird! reissbaker 8 days ago It was natively trained in FP4. Probably both to reduce VRAM usage at inference time (fits on a single H100), and to allow better utilization of B200s (which are especially fast for FP4). irthomasthomas 8 days ago Interesting, thanks. I didn't know you could even train at FP4 on H100s 1 reply →
irthomasthomas 8 days ago Oh, I didn't know that. Weird! reissbaker 8 days ago It was natively trained in FP4. Probably both to reduce VRAM usage at inference time (fits on a single H100), and to allow better utilization of B200s (which are especially fast for FP4). irthomasthomas 8 days ago Interesting, thanks. I didn't know you could even train at FP4 on H100s 1 reply →
reissbaker 8 days ago It was natively trained in FP4. Probably both to reduce VRAM usage at inference time (fits on a single H100), and to allow better utilization of B200s (which are especially fast for FP4). irthomasthomas 8 days ago Interesting, thanks. I didn't know you could even train at FP4 on H100s 1 reply →
irthomasthomas 8 days ago Interesting, thanks. I didn't know you could even train at FP4 on H100s 1 reply →
Oh, I didn't know that. Weird!
It was natively trained in FP4. Probably both to reduce VRAM usage at inference time (fits on a single H100), and to allow better utilization of B200s (which are especially fast for FP4).
Interesting, thanks. I didn't know you could even train at FP4 on H100s
1 reply →