zlacker

Hi everyone yes, I left OpenAI yesterday

submitted by mfigui+(OP) on 2024-02-14 03:08:18 | 769 points 444 comments
[view article] [source] [go to bottom]

NOTE: showing posts with links only show all posts
6. stygia+hMaF2[view] [source] 2024-02-14 02:13:37
>>mfigui+(OP)
It seems like he (re)joined OpenAI almost exactly 1 year ago: https://twitter.com/karpathy/status/1623476659369443328
◧◩
22. george+pPaF2[view] [source] [discussion] 2024-02-14 02:43:19
>>dontre+ROaF2
> Relative to his level of fame, his actual level of contribution as far as pushing forward AI, I’m not so sure about.

Are you sure about your perspective?

https://scholar.google.com/citations?view_op=view_citation&h...

◧◩◪
25. the_ar+jRaF2[view] [source] [discussion] 2024-02-14 02:59:40
>>george+pPaF2
May be this is what you wanted to share? https://scholar.google.com/citations?user=l8WuQJgAAAAJ&hl=en
◧◩◪
32. Imnimo+o[view] [source] [discussion] 2024-02-14 03:11:04
>>skybri+9SaF2
The most recent is this, which I believe was made after he left Tesla:

https://github.com/karpathy/nanoGPT

And it's accompanying video series:

https://karpathy.ai/zero-to-hero.html

Another example (although I honestly don't remember if he made this one between jobs) is: https://github.com/karpathy/micrograd

◧◩◪
33. joss82+y[view] [source] [discussion] 2024-02-14 03:12:09
>>skybri+9SaF2
Neural Networks: from zero to hero

https://karpathy.ai/zero-to-hero.html

48. gnabgi+p4[view] [source] 2024-02-14 03:46:54
>>mfigui+(OP)
Related: "Andrej Karpathy Departs OpenAI"[0] (159 points, 2 hours ago, 71 comments)

[0]: >>39365288

◧◩
58. 0xcde4+R6[view] [source] [discussion] 2024-02-14 04:09:09
>>Prolly+H1
I don't mean to detract from your point (if anything, I suppose I'm obliquely supporting it), but I feel compelled to say that it's really weird to see Carver Mead cited in the context of "employee at Xerox PARC", because I mostly know him as one half of "Mead/Conway", i.e. the duo who arguably supplied the computational (dare I say "algorithmic"?) rocket fuel for the unbelievably wild progress of chips in the 1990s [1] [2].

[1] https://en.wikipedia.org/wiki/Mead%E2%80%93Conway_VLSI_chip_...

[2] https://en.wikipedia.org/wiki/Lynn_Conway

◧◩◪
67. Prolly+q8[view] [source] [discussion] 2024-02-14 04:22:31
>>0xcde4+R6
The textbook they wrote together was while both were collaborating at PARC (Mead was at CalTech, then, too); they wrote it to add credibility to their VLSI theories, which at the time most experts believed would lead to thermal runaway (i.e. not stable, long-term, to pack transistors densely).

https://en.wikipedia.org/wiki/Carver_Mead

Learning about the interconnectedness of all this historic intellectual "brain theft," keeps me excited for an AGI-future, post-copyright/IP. What are we going to accomplish [globally] when you can't just own brilliant ideas?!

◧◩◪
69. squigz+V8[view] [source] [discussion] 2024-02-14 04:28:43
>>KerrAv+O7
https://en.wikipedia.org/wiki/Bee_line
◧◩◪
71. weinzi+h9[view] [source] [discussion] 2024-02-14 04:33:01
>>skybri+9SaF2
Andrej Karpathy is badmephisto, a name you might have heard of if you're into cubing.

http://badmephisto.com/

76. nutanc+fa[view] [source] 2024-02-14 04:41:59
>>mfigui+(OP)
He quit to build a blogging platform.

https://x.com/karpathy/status/1751350002281300461?s=20

◧◩◪◨⬒
107. Prolly+Te[view] [source] [discussion] 2024-02-14 05:33:54
>>wolver+dc
>Very few people knew they even existed

Does not mean they did not exist. See citations, below:

https://en.wikipedia.org/wiki/Laser_printing (see 2nd intro paragraph)

https://en.wikipedia.org/wiki/History_of_email (see 3rd intro paragraph)

◧◩◪◨
116. sho+Uf[view] [source] [discussion] 2024-02-14 05:47:20
>>pb7+Af
here was one of them https://gemini.google.com/share/fde31202b221?hl=en

edit: as pointed out, this was indeed a pretty esoteric example. But the rest of my attempts were hardly better, if they had a response at all.

◧◩
119. magogh+Ag[view] [source] [discussion] 2024-02-14 05:56:26
>>dontre+ROaF2
In 2015 he wrote this blog post about "The Unreasonable Effectiveness of Recurrent Neural Networks": https://karpathy.github.io/2015/05/21/rnn-effectiveness/

That blog post inspired Alec Radford at Open AI to do the research that produced the "Unsupervised sentiment neuron": https://openai.com/research/unsupervised-sentiment-neuron

Open AI decided to see what happened if they scaled up that model by leveraging the new Transformer architecture invented at Google, and they created something called GPT: https://cdn.openai.com/research-covers/language-unsupervised...

◧◩◪◨⬒
127. threes+ci[view] [source] [discussion] 2024-02-14 06:20:57
>>trhway+oa
> which they get from cameras like one can get it from lidar

LiDAR directly measures the distance to objects. What Tesla is doing is inferring it from two cameras.

There has been plenty of research to date [1] that LiDAR + Vision is significantly better than Vision Only especially under edge case conditions e.g. night, inclement weather when determining object bounding boxes.

[1] https://iopscience.iop.org/article/10.1088/1742-6596/2093/1/...

136. ragebo+2k[view] [source] 2024-02-14 06:42:25
>>mfigui+(OP)
In his free time, I hope he writes some more fiction, I really liked https://karpathy.github.io/2015/11/14/ai/
◧◩◪
148. passio+um[view] [source] [discussion] 2024-02-14 07:14:20
>>bobthe+bc
Become the next 3blue1brown. He has inspired many.

Here's a gem of educator. Check out his other videos.

https://www.youtube.com/watch?v=dhYqflvJMXc

◧◩◪◨
149. jasmat+Am[view] [source] [discussion] 2024-02-14 07:15:03
>>weinzi+h9
I haven't been that surprised by something in a long time. Wow that is crazy. I made a little unfinished 3d Rubik's Cube site for fun a while back and the about section includes a link to his channel and some other older cubing channels. https://rubie-cubie.vercel.app/
◧◩
166. daniel+Mr[view] [source] [discussion] 2024-02-14 08:13:02
>>skepti+FOaF2
Googlers are wishing OpenAI could vanish as it makes them look like the IBM-lookalike they are.

Here are some hilarious highlights: https://twitter.com/Suhail/status/1757573182138290284

◧◩◪
197. wwilim+AB[view] [source] [discussion] 2024-02-14 10:02:35
>>skybri+9SaF2
My master's was in Convolutional NNs for language processing. I had zero prior knowledge and my advisor recommended I watch Karpathy's lectures[1] to get up to speed

[1] https://youtube.com/playlist?list=PLkt2uSq6rBVctENoVBg1TpCC7...

◧◩◪◨⬒⬓⬔
223. nindal+NG[view] [source] [discussion] 2024-02-14 11:14:15
>>LightB+IE
> They're not brain surgeons or rocket scientists

Yeah, voters don't want to pay MPs more. Yet when voters are asked, they want highly intelligent, motivated people. They want them to have technical expertise, which means time spent in higher education. Then they want them to work a full time job in Parliament during the week, but also be open to constituency concerns on the weekend. And once all of this is pointed out, voters concede that maybe MPs deserve to be paid on par with professionals like doctors. (It's a different matter that UK doctors are underpaid).

> But I believe the average gravy train bumps this up 3X with extras.

Citation needed. They're on a shorter leash now with expenses. Don't go citing one or two bad apples either, show us what the median MP claims as expenses. According to you, it should be around £170k a year.

In general, politicians and their aides in the UK are underpaid. Most capable people find they're better off working in Canary Wharf or elsewhere in London. An example is the head of economic policy for the Labour Party earning £50k while writing policy for a £2 trn economy. (https://www.economist.com/britain/2023/01/19/british-politic...)

◧◩◪◨⬒
228. TeMPOr+PH[view] [source] [discussion] 2024-02-14 11:24:39
>>bamboo+jD
The author of `aider` - an OSS GPT-powered coding assistant - is on HN, and says[0] he has benchmarks showing gradual decline in quality of GPT-4-Turbo, especially wrt. "lazy coding" - i.e. actually completing a coding request, vs. peppering it with " ... write this yourself ... " comments.

That on top of my own experiences, and heaps of anecdotes over the last year.

> How would they honestly be getting worse?

The models behind GPT-4 (which is rumored to be a mixture model)? Tuning, RLHF (which has long been demonstrated to dumb the model down). The GPT-4, as in the thing that produces responses you get through API? Caching, load-balancing, whatever other tricks they do to keep the costs down and availability up, to cope with the growth of the number of requests.

--

[0] - >>39361705

◧◩◪
235. anoope+OJ[view] [source] [discussion] 2024-02-14 11:47:15
>>skybri+9SaF2
And he was teaching CS231n in Stanford in 2016

https://www.youtube.com/watch?v=NfnWJUyUJYU&list=PLkt2uSq6rB...

◧◩◪◨⬒
238. bayind+8K[view] [source] [discussion] 2024-02-14 11:48:43
>>inglor+kJ
On the other hand, Israel is using AI to generate their bombing targets and pound Gaza strip with bombs non-stop [0].

And, according to UN, Turkey has used AI powered, autonomous littering drones to hit military convoys in Libya [1].

Regardless of us vs. them, AI shouldn't be a part of warfare, IMHO.

[0]: https://www.theguardian.com/world/2023/dec/01/the-gospel-how...

[1]: https://www.voanews.com/a/africa_possible-first-use-ai-armed...

◧◩◪◨
249. Copenj+JO[view] [source] [discussion] 2024-02-14 12:30:49
>>weinzi+h9
10 years ago: https://youtu.be/WhPjlnWbtS8?feature=shared&t=359
◧◩◪◨⬒⬓⬔
254. bayind+QQ[view] [source] [discussion] 2024-02-14 12:47:01
>>dizhn+EM
Voice of America generally employs country's nationals for their reporting. There are some other resources:

    - NPR: https://www.npr.org/2021/06/01/1002196245/a-u-n-report-suggests-libya-saw-the-first-battlefield-killing-by-an-autonomous-d
    - Lieber Institute: https://lieber.westpoint.edu/kargu-2-autonomous-attack-drone-legal-ethical/
    - ICRC: https://casebook.icrc.org/case-study/libya-use-lethal-autonomous-weapon-systems
    - UN report itself (Search for Kargu): https://undocs.org/Home/Mobile?FinalSymbol=S%2F2021%2F229&Language=E&DeviceType=Desktop&LangRequested=False
    - Kargu itself: https://www.stm.com.tr/en/kargu-autonomous-tactical-multi-rotor-attack-uav
From my experience, Turkish military doesn't like to talk about all the things they have.
◧◩◪◨
275. buggle+sX[view] [source] [discussion] 2024-02-14 13:41:18
>>Keyfra+AT
The solution, at least for GPT-4, is to ask it to first draft a software spec for whatever you want it to implement and then write the code based on the spec. There are a bunch of examples here:

https://github.com/mckaywrigley/prompts

◧◩
291. CamelC+g21[view] [source] [discussion] 2024-02-14 14:10:01
>>option+dOaF2
Huh, I just read this exact tweet.

https://x.com/ns123abc/status/1757595970911556012

◧◩◪◨
296. magogh+251[view] [source] [discussion] 2024-02-14 14:24:52
>>arugul+Tq
I tried to find the where I heard that Radford was inspired by that blog post, but the closest thing I found is that in the "Sentiment Neuron" paper (Learning to Generate Reviews and Discovering Sentiment: https://arxiv.org/pdf/1704.01444.pdf), in the "Discussion and Future Work" section they mention this Karpathy paper from 2015: Visualizing and Understanding Recurrent Networks https://arxiv.org/abs/1506.02078
◧◩◪◨⬒⬓⬔
329. jorvi+wq1[view] [source] [discussion] 2024-02-14 15:58:45
>>bigyik+Z01
That’s kind of the point, you won’t be able to due to the algorithm.

I can give you something analogous though: I’m a big fan of old school east coast hip-hop. You have the established mainline artists from back then (“Nas”, “Jay-Z”, “Big L”, etc), then you have a the established underground artists (say, “Lord Finesse” or “Kool G Rap”), and then you have the really really underground guys like “Mr. Low Kash ‘n Da Shady Bunch”, “Superscientifiku”, “Punk Barbarians”, “Harlekinz”, etc.

A lot of those in that third “tier” are every bit as good as the second tier. And both tiers contain a lot of artists that could hit the quality point of the mainline artists, they just never had access to the producer and studio time that the mainline did.

I know these artists because I love going digging for the next hidden gem. Spotify recommended me perhaps one or two of all the super-underground guys.

Ironically more West-coast style, but here is a great example (explicit!): https://youtu.be/BUwJMVKSMtY?t=129

Dude could’ve measured up to the best of the west coast. Spotify monthly listener count? 891.

Algorithms are sadly win-more.

Now I’m just silently hoping a math nerd will feel inclined to share their hidden math channel gems :+)

◧◩◪◨⬒⬓⬔
364. mikhai+WK1[view] [source] [discussion] 2024-02-14 17:42:34
>>bigyik+Z01
3blue1brown runs Summer of Math competitions to highlight other creative math videos. Many, but not all, use the same 3b1b 'manim' animation software, so they often have the same look'n'feel. Here are the results from 2022, and the huge YT playlist:

https://www.3blue1brown.com/blog/some2

https://www.youtube.com/playlist?list=PLnQX-jgAF5pTZXPiD8ciE...

◧◩◪
365. mandee+bM1[view] [source] [discussion] 2024-02-14 17:49:04
>>next_x+l3
> LLM based OS

Isn't that is what Rabbit R1 is? https://www.rabbit.tech/

◧◩◪◨⬒⬓⬔⧯▣
380. jorvi+7h2[view] [source] [discussion] 2024-02-14 20:15:12
>>_tk_+4T1
(I realize know I dislike using the descriptor "tier", as it implies some sort of ranking. Perhaps "layer" would have been better, but I'll stick with it for now)

For both tier 2 and tier 3 its basically the same process. This is for Spotify btw, I have no idea how different the workflow would be for something like Apple Music.

Say the genre you want to dig around in is Hip-Hop. You are aware of Eminem and Mac Miller, and vaguely aware of a guy named Nas. By intuition you'd probably already be able to tell that Nas is more at the edge among the mainline artists.

You click on "Nas", and scroll down to Fans also like. Right now, for "Nas", it is showing "Mobb Deep", "Mos Def", "Rakim", "Big L", "Wu-Tang Clan", "Gang Starr", "Ghostface Killah", "Method Man" and "Common".

This is a mix T1 and T2. "Wu-Tang"s in there along with assorted members, but some of the other artists are much lesser known quantities.

Its a bit hard for me to decide what a Hip-Hop layman would consider the most unknown name here, but I'd venture it'd be "Big L". We click on him, do the same thing. Now we're really getting somewhere, with guys like "Inspectah Deck" and "Smif-n-Wessun". Click, dig, we get a bunch of names amongst which "Lord Finesse" stands out. The Show more at the end of Fans Like is also invaluable.

In total the dig order for me to get to the very bottom of the undeground is "Nas" > "Big L" > "Smif-n-Wessun" > "Lord Finesse" > "Channel Live" > "Ed OG & Da Bulldogs" > "Trends of Culture" > "Brokin English Klik" (358 monthly listeners).

I wouldn't consider each of those going a tier (layer) deeper. As a guy who knows waaay too much about Hip-Hop, I'd separate them into:

- T1: "Nas", "Big L"

- T2 "Smif-n-Wessun", "Lord Finesse"

- T3 "Channel Live", "Ed OG & Da Bulldogs", "Trends of Culture", "Brokin English Klik"

Perhaps "Brokin English Klik" should be in its own T4 and 3 tiers lacks the fidelity to be necessarily accurate. Not sure.

A little shortcut would be using "The Edge of $Genre" playlists. They're the pair playlists to "The Sound of $Genre" (broad slice) and "The Pulse of $Genre" (most popular) generated via everynoise.com, although as that guy got fired from Spotify its up in the air how long those will keep working.

Edit: oh, and if you run into a playlist that caters to that deep underground (in my case, that was "90's Tapes"*), that's worth its bytes in gold.

*https://open.spotify.com/playlist/2H0rNGEBShvHSGebM2m37c?si=...

◧◩◪◨⬒⬓⬔
414. plutok+cB3[view] [source] [discussion] 2024-02-15 06:34:19
>>bigyik+Z01
https://www.youtube.com/@Reducible/videos
◧◩
424. CarlsJ+xn5[view] [source] [discussion] 2024-02-15 18:43:59
>>skepti+FOaF2
Genuinely curious if the news today about Sora has changed your opinion at all https://openai.com/sora
◧◩
440. wrsh07+kXd[view] [source] [discussion] 2024-02-18 15:02:15
>>d--b+OE
He was building small, efficient versions of gpt 3 in a video series

You can check his GitHub: https://github.com/karpathy

[go to top]