zlacker

[parent] [thread] 3 comments
1. Mockap+(OP)[view] [source] 2022-05-23 21:27:31
Transformers are parallelize-able, right? What’s stopping a large group of people from pooling their compute power together and working towards something like this? IIRC there were some crypto projects a while back that we’re trying to create something similar (golem?)
replies(3): >>visarg+71 >>joshcr+x4 >>sineno+Qz
2. visarg+71[view] [source] 2022-05-23 21:33:00
>>Mockap+(OP)
There are the Eleuther.ai and BigScience projects working on public foundation models. They have a few releases already and currently training GPT-3 sized models.
3. joshcr+x4[view] [source] 2022-05-23 21:52:44
>>Mockap+(OP)
There are people working on reproducing the models, see here for Dall-E 2 for example: https://github.com/lucidrains/DALLE2-pytorch

It's often not worth it to decentralize the computation of the trained model though but it's not hard to get donated cycles and groups are working on it. Don't fret because Google isn't releasing the API/code. They released the paper and that's all you need.

4. sineno+Qz[view] [source] 2022-05-24 01:55:05
>>Mockap+(OP)
You really need a decent infiniband-linked cluster to train large models.
[go to top]