zlacker

[parent] [thread] 2 comments
1. ianbut+(OP)[view] [source] 2026-02-03 20:44:06
All models released from those providers go through stages of post training too, none of the models you interact with go from pre-training to release. An example of the post training pipeline is tool calling, that is to my understanding a part of post training and not pre training in general.

I can't speak to what the exact split is or what is a part of post training versus pre training at various labs but I am exceedingly confident all labs post train for effectiveness in specific domains.

replies(1): >>Der_Ei+z
2. Der_Ei+z[view] [source] 2026-02-03 20:46:55
>>ianbut+(OP)
I did not claim that post training doesn't happen on these models, and you are being extremely patronizing (I publish quite a bit of research on LLMs at top conferences).

I claimed that OpenAI overindexed on getting away with aggressive post-training on old pre-training checkpoints. Gemini / Anthropic correctly realized that new pre-training checkpoints need to happen to get the best gains in their latest model releases (which get post-trained too).

replies(1): >>ianbut+Iy
◧◩
3. ianbut+Iy[view] [source] [discussion] 2026-02-03 23:48:01
>>Der_Ei+z
If you read that as patronizing that says more about you than me personally, I have no idea who you are so your own insecurity at what is a rather unloaded explanation perplexes me.
[go to top]