For one, I'm not sure Sam Altman will tolerate MS bureaucracy for very long.
But secondly, the new MS-AI entity can't presumably just take from OpenAI what they did there, they need to make it again.
This takes a lot of resources (that MS has) but also a lot of time to provide feedback to the models; also, copyright issues regarding source materials are more sensitive today, and people are more attuned to them: Microsoft will have a harder time playing fast and lose with that today, than OpenAI 8 years ago.
Or, Sam at MS becomes OpenAI biggest customer? But in that case, what are all those researchers and top scientists that followed him there, going to do?
Interesting times in any case.
The dataset is more challenging, but here msft can help - since they have bing and github as well. So they might be able to make few shortcuts here.
The most time consuming part is compute, but here again msft has the compute.
Will they beat chat-gpt 4 in a year? Guess no. But they will come very close to it and maybe it would not matter that much if you focus on the product.
Part of me thinks that Nadella, having already demonstrated his mastery over all his competitor CEOs with one deft move after another over the past few years, took this on because he needed a new challenge.
I'd wager Altman will either get sidelined and pushed out, or become Nadella's successor, over the course of the next decade or so.
It's an interesting time!
What I meant is, most likely assuming that you are using pytorch / jax you could code down the model pretty fast. Just compare it to llama, sure it is far behind, but the llama model is under 1000 lines of code and pretty good.
There is tons of work, for the training, infra, preparing the data and so on. That would result guess in millions lines of code. But the core ideas and the model are likely thin I would argue. So that is my point.