It’s a bit tragic that Ilya and company achieved the exact opposite of what they intended apparently, by driving those they attempted to slow down into the arms of people with more money and less morals. Well.
If they didn’t fire him, Altman will just continue to run hog wild over their charter. In that sense they lose either way.
At least this way, OpenAI can continue to operate independently instead of being Microsoft’s zombie vassal company with their mole Altman pulling the strings.
It seems like people forget that it was the investors’ money that made all this possible in the first place.
There are probably load so ways you can make language models with 100M parameters more efficient, but most of them won't scale to models with 100B parameters.
IIRC there is a bit of a phase transition that happens around 7B parameters where the distribution of activations changes qualitatively.
Anthropic have interpretability papers where their method does not work for 'small' models (with ~5B parameters) but works great for models with >50B parameters.