zlacker

[parent] [thread] 2 comments
1. Michae+(OP)[view] [source] 2023-11-20 03:49:16
i suspect that 'AI safety' means a chatbot that is not hallucinating/making things up. Ilya Sutskever may want some sort of hybrid system, where the output of the LLM gets vetted by a second system, so as to minimize instances of hallucinations, whereas Sam Altman says 'screw it, lets make an even bigger LLM and just push it'.

Is that right?

Don't know if Altman or Sutskever is right, there seems to be a kind of arms race between the companies. OpenAI may be past the point where they can try out a radically different system, due to competition in the space. Maybe trying out new approaches could only work in a new company, who knows?

replies(1): >>pixl97+K7
2. pixl97+K7[view] [source] 2023-11-20 05:08:37
>>Michae+(OP)
AI Safety means a lot of different things.

Inward safety for people means their vulnerabilities are not exposed and/or open to attack. Outward safety means they are not attacking others and looking out for the general wellbeing of others. We have a lot of different social constructs with other people of which we attempt to keep this in balance. It doesn't work out for everyone, but in general it's somewhat stable.

What does it mean to be safe if you're not at threat of being attacked and harm? What does attacking others mean if doing so is meaningless, just a calculation? This goes from everything from telling someone to kill themselves (it's just words) to issuing a set of commands to external devices with real world effects (print ebola virus or launch nuclear weapon). The concern here is the AI of the future will be extraordinarily powerful yet very risky when it comes to making decisions that could harm others.

replies(1): >>Michae+Br
◧◩
3. Michae+Br[view] [source] [discussion] 2023-11-20 07:09:52
>>pixl97+K7
yes, still it boils down to the question of 'how well is the output of the chat bot aligned with reality' ? if you want to automate this, then you will likely need some system that is kind of censoring the output of the LLM, and that system should have a better model of what is real.
[go to top]