zlacker

[return to ""]
1. dontre+(OP)[view] [source] 2024-02-14 02:37:27
>>mfigui+A3
Unpopular opinion… but IMO almost all of Karpathy’s fame an influence come from being an incredible educator and communicator.

Relative to his level of fame, his actual level of contribution as far as pushing forward AI, I’m not so sure about.

I deeply appreciate his educational content and I’m glad that it has led to a way for him to gain influence and sustain a career. Hopefully he’s rich enough from that that he can focus 100% on educational stuff!

2. mfigui+A3[view] [source] 2024-02-14 03:08:18
â—§
3. magogh+ak[view] [source] 2024-02-14 05:56:26
>>dontre+(OP)
In 2015 he wrote this blog post about "The Unreasonable Effectiveness of Recurrent Neural Networks": https://karpathy.github.io/2015/05/21/rnn-effectiveness/

That blog post inspired Alec Radford at Open AI to do the research that produced the "Unsupervised sentiment neuron": https://openai.com/research/unsupervised-sentiment-neuron

Open AI decided to see what happened if they scaled up that model by leveraging the new Transformer architecture invented at Google, and they created something called GPT: https://cdn.openai.com/research-covers/language-unsupervised...

â—§â—©
4. arugul+tu[view] [source] 2024-02-14 08:03:58
>>magogh+ak
Is it stated somewhere that Radford was inspired by that blog post?
â—§â—©â—ª
5. magogh+C81[view] [source] 2024-02-14 14:24:52
>>arugul+tu
I tried to find the where I heard that Radford was inspired by that blog post, but the closest thing I found is that in the "Sentiment Neuron" paper (Learning to Generate Reviews and Discovering Sentiment: https://arxiv.org/pdf/1704.01444.pdf), in the "Discussion and Future Work" section they mention this Karpathy paper from 2015: Visualizing and Understanding Recurrent Networks https://arxiv.org/abs/1506.02078
[go to top]