zlacker

[parent] [thread] 25 comments
1. mgdev+(OP)[view] [source] 2024-05-17 15:44:40
Yes, it's valuable to have a small research team who focuses on R&D outside the production loop.

But when you give them a larger remit, and structure teams with some owning "value" and and others essentially owning "risk", the risk teams tend to attract navel-gazers and/or coasters. They wield their authority like a whip without regard for business value.

The problem is the incentives tend to be totally misaligned. Instead the team that ships the "value" also needs to own their own risk management - metrics and counter metrics - with management holding them accountable for striking the balance.

replies(8): >>seanhu+91 >>Hasu+j1 >>23B1+q1 >>germin+N2 >>reduce+T5 >>davidi+69 >>encode+Vb >>6gvONx+Cq
2. seanhu+91[view] [source] 2024-05-17 15:51:06
>>mgdev+(OP)
Completely agree with this. Everyone doing AI has to do AI that is both valuable and responsible. You can't have an org structure where the valuable AI and responsible AI teams are in some kind of war against each other. It will just never work well.[1]

Imagine if vehicle manufacturers[2] split their design and R&D teams into a "make the thing go" team and a "don't kill the passengers" team. Literally noone would think that arrangement made sense.

I can totally see when we are at a state of significant maturity of both AI and AI regulation that you have a special part of your legal team that are specialised in legal/regulatory compliance issues around AI just like companies tend to have specialised data privacy compliance experts. But we're not there yet.

[1] If you're serious about long-term AI risk and alignment research, sponsor some independent academic research that gets published. That way it's arms-length and genuinely credible.

[2] If you like you can maybe mentally exclude Boeing in this.

replies(2): >>mgdev+k3 >>doktri+a7
3. Hasu+j1[view] [source] 2024-05-17 15:52:09
>>mgdev+(OP)
The purpose of the risk team at OpenAI was to prevent the destruction of humanity.

I think you definitely want people who have that responsibility to "wield their authority like a whip without regard for business value".

Now, whether you buy OpenAI's hype about the potential danger (and value) of their products, that's up to you, but when the company says, "We're getting rid of the team that makes sure we don't kill everyone", there is a message being sent. Whether it's "We don't really think our technology is that dangerous (and therefore valuable)" or "We don't really care if we accidentally kill everyone", it's not a good message.

replies(2): >>square+v3 >>mgdev+b5
4. 23B1+q1[view] [source] 2024-05-17 15:52:46
>>mgdev+(OP)
No, this just puts the fox in the henhouse. Systems of checks-and-balances between independent entities exist for a reason.

Without them internally, it'll just fall to regulators, which of course is what shareholders want; to privatize upside and socialize downside.

replies(1): >>mgdev+H2
◧◩
5. mgdev+H2[view] [source] [discussion] 2024-05-17 15:59:48
>>23B1+q1
Agree that you need checks and balances, but there are better and worse systems.

As someone who has scaled orgs from tens to thousands of engineers, I can tell you: you need value teams to own their own risk.

A small, central R&D team may work with management to set the bar, but they can't be responsible for mitigating the risk on the ground - and they shouldn't be led to believe that that is their job. It never works, and creates bad team dynamics. Either the central team goes too far, or they feel ignored. (See: security, compliance.)

6. germin+N2[view] [source] 2024-05-17 16:00:25
>>mgdev+(OP)
In the general case, I mostly agree, but it cracks me up that this is the prevailing attitude when it comes to our industry; but when we see police departments or government agencies trying to follow the same playbook, we immediately point out how that's laughable and doesn't result in real accountability.

In this specific case, though, Sam Altman's narrative is that they created an existential risk to humanity and that the access to it needs to be restricted for others. So which is it?

replies(1): >>nprate+C4
◧◩
7. mgdev+k3[view] [source] [discussion] 2024-05-17 16:04:20
>>seanhu+91
> Imagine if vehicle manufacturers[2] split their design and R&D teams into a "make the thing go" team and a "don't kill the passengers" team.

Rocket caskets. Can't kill someone who is already dead!

◧◩
8. square+v3[view] [source] [discussion] 2024-05-17 16:04:56
>>Hasu+j1
> but when the company says, "We're getting rid of the team that makes sure we don't kill everyone", there is a message being sent

Hard not to imagine a pattern if one considers what they did a few months ago:

https://www.cnbc.com/2024/01/16/openai-quietly-removes-ban-o...

replies(1): >>mnky98+9l1
◧◩
9. nprate+C4[view] [source] [discussion] 2024-05-17 16:12:02
>>germin+N2
Neither. It was pure hype to get free column inches.

Anyone who's used their AI and discovered how it ignores instructions and makes things up isn't going to honestly believe it poses an existential threat any time soon. Now they're big enough they can end that charade.

replies(1): >>saulpw+D5
◧◩
10. mgdev+b5[view] [source] [discussion] 2024-05-17 16:14:52
>>Hasu+j1
> The purpose of the risk team at OpenAI was to prevent the destruction of humanity.

Yeah, the problem (in this outsider's opinion) is that that charter is so ill-defined that it's practically useless, which in turn means that any sufficiently loud voice can apply it to anything. It's practically begging to be used as a whip.

> I think you definitely want people who have that responsibility to "wield their authority like a whip without regard for business value".

No, because without thinking of value, there is no enterprise, and then your mission is impossible. Essentially, it creates an incentive where the best outcome is to destroy the company. And, hey hey, that's kinda what almost happened.

> Whether it's "We don't really think our technology is that dangerous (and therefore valuable)" or "We don't really care if we accidentally kill everyone", it's not a good message.

I don't think it has to be so black-and-white as this. Meta, Microsoft, and Google did the same thing. Instead, those functions have been integrated more closely into the value teams. And I can't imagine Amazon or Oracle ever having those teams in the first place. They likely all realized the same thing: those teams add huge drag without adding measurable business value.

And yes, there are ways to measure the business value of risk management, and weigh against upside value to decide the correct course of actions - it's just that most of those teams in big tech don't actually take a formal "risk management" approach. Instead they pontificate or copy and enforce.

◧◩◪
11. saulpw+D5[view] [source] [discussion] 2024-05-17 16:17:37
>>nprate+C4
Ignoring instructions and making things up may very well be an existential threat. Just not of the SkyNet variety.
replies(1): >>nprate+ob
12. reduce+T5[view] [source] 2024-05-17 16:18:53
>>mgdev+(OP)
Did you read Jan Leike's resignation? https://x.com/janleike/status/1791498174659715494

I hope others see that there are two extremely intelligent sides, but one has mega $$ to earn and the other is pleading that there are dangers ahead and not to follow the money and fame.

This is climate change and oil companies all over again, and just like then and now, oil companies are winning.

Fundamentally, many people are the first stage, denial. Staring down our current trajectory of AGI is one of the darkest realities to imagine and that is not pleasant to grapple with.

replies(2): >>mgdev+V8 >>swatco+v9
◧◩
13. doktri+a7[view] [source] [discussion] 2024-05-17 16:26:06
>>seanhu+91
1. Audit / evaluation / quality assurance teams exist across multiple verticals from multinationals to government, and cannot reliably function when overly subservient to the production or “value creating” side

2. Boeing is a good and timely example of the consequences of said internal checks and balances collapsing under “value creation” pressure. That was a catastrophic failure which still can’t reasonably be compared to the downside of misaligned AI.

replies(1): >>seanhu+Z7
◧◩◪
14. seanhu+Z7[view] [source] [discussion] 2024-05-17 16:30:22
>>doktri+a7
I agree with you on both points, but they have QA which is 1. The long-term risk team was more of a research/futurology/navel-gazing entity rather than a qa/audit function. I would say if you have any possible safety/alignment test that you can feasibly run it should be part of the CI/CD pipline and be run during training also. That's not what that group was doing.
replies(1): >>doktri+WN1
◧◩
15. mgdev+V8[view] [source] [discussion] 2024-05-17 16:34:54
>>reduce+T5
I hadn't until now, but it's the other failure mode I mentioned in another fork [1] of this thread:

> A small, central R&D team may work with management to set the bar, but they can't be responsible for mitigating the risk on the ground - and they shouldn't be led to believe that that is their job. It never works, and creates bad team dynamics. Either the central team goes too far, or they feel ignored. (See: security, compliance.)

[1]: >>40391283

16. davidi+69[view] [source] 2024-05-17 16:35:58
>>mgdev+(OP)
So what you're saying is OpenAI can't align two teams internally but they want to align a super intelligence.
replies(1): >>mgdev+gd
◧◩
17. swatco+v9[view] [source] [discussion] 2024-05-17 16:38:38
>>reduce+T5
The alternate take is that the ceiling is proving to be lower than originally hoped, and that one team of startup folk are looking to squeeze out a business model that satisfies their original investors and the other team of passionate researchers are ready to go find their next bold research opportunities elsewhere.
replies(2): >>mgdev+uc >>reduce+We
◧◩◪◨
18. nprate+ob[view] [source] [discussion] 2024-05-17 16:49:26
>>saulpw+D5
Of course, but stupidity and lying always potentially is. It doesn't need to be technical. That's a whole different conversation.
19. encode+Vb[view] [source] 2024-05-17 16:53:09
>>mgdev+(OP)
Here’s the former team leaders take. Says they couldn’t get compute resources to do their job.

https://x.com/janleike/status/1791498174659715494

replies(1): >>petre+vv1
◧◩◪
20. mgdev+uc[view] [source] [discussion] 2024-05-17 16:56:27
>>swatco+v9
I like this take best. :)

Most folks who are part of the startup, don't care to stick around for the scale. Different phases of the company tend to attract different types of people.

◧◩
21. mgdev+gd[view] [source] [discussion] 2024-05-17 17:01:15
>>davidi+69
I think "aligning super intelligence" is a nothingburger of a goal, for exactly that reason. It's not a problem that's unique to OpenAI.

The reason you can't "align" AI is because we, as humans on the planet, aren't universally aligned on what "aligned" means.

At best you can align to a particular group of people (a company, a town, a state, a country). But "global alignment" in almost any context just devolves into war or authoritarianism (virtual or actual).

◧◩◪
22. reduce+We[view] [source] [discussion] 2024-05-17 17:09:44
>>swatco+v9
> passionate researchers are ready to go find their next bold research opportunities elsewhere

Not supported by Ilya agreeing with board to fire Sam Altman.

I also think you'll struggle to find a majority of people thinking AI research's "ceiling is proving to be lower than originally hoped", what with 4o, SORA, GPT5 all coming and it's only been 1.5 years since ChatGPT.

23. 6gvONx+Cq[view] [source] 2024-05-17 18:20:14
>>mgdev+(OP)
Speaking of incentives, the way your suggestion tends to work out is that you make a team own both value and risk, then incentivize them only for value, which works out predictably.
◧◩◪
24. mnky98+9l1[view] [source] [discussion] 2024-05-18 03:27:04
>>square+v3
Maybe the message is, these ai ain't going to take over the world.
◧◩
25. petre+vv1[view] [source] [discussion] 2024-05-18 06:37:33
>>encode+Vb
Well, Altman did try to get potential investors to shell out $7T for "AI chips".
◧◩◪◨
26. doktri+WN1[view] [source] [discussion] 2024-05-18 11:22:33
>>seanhu+Z7
That’s quite the narrow goalpost you’ve set up. What happens if a problem can’t be expressed as a Jenkins pipeline operation?
[go to top]