zlacker

[parent] [thread] 1 comments
1. salmon+(OP)[view] [source] 2023-05-16 16:00:47
> From its training data GPT-7 might notice

> But its "aligned" so might understand

> Using this information it decides to hack

I think you're anthropomorphizing LLM's too much here. If we assume that there's a AGI-esque AI, then of course we should be worried about an AGI-esque AI. But I see no reason to think that's the case.

replies(1): >>HDThor+Yb1
2. HDThor+Yb1[view] [source] 2023-05-16 21:55:58
>>salmon+(OP)
The whole issue with near term alignment is that people will anthropomorphize AI. That’s what it being unaligned means, it’s treated like a responsible person when it in fact is not. I don’t think it’s hard at all to think of a scenario where a dumb as rocks agentic ai gives itself the task of accumulating more power since its training data says having power helps solve problems. From there it again doesn’t have to be anything other than a stochastic parrot to order people to do horrible things.
[go to top]