zlacker

[return to "A Timeline of the OpenAI Board"]
1. upward+A6[view] [source] 2023-11-19 08:44:50
>>prawn+(OP)
> they asked: who on earth are Tasha McCauley and Helen Toner?

As a prominent researcher in AI safety (I discovered prompt injection) I should explain that Helen Toner is a big name in the AI safety community - she’s one of the top 20 most respected people in our community, like Rohin Shah.

The “who on earth” question is a good question about Tasha. But grouping Helen in with Tasha is just sexist. By analogy, Tasha is like Kimbal Musk, whereas Helen is like Tom Mueller.

Tasha seems unqualified but Helen is extremely qualified. Grouping them together is sexist and wrong.

◧◩
2. hhjink+Nj[view] [source] 2023-11-19 10:48:30
>>upward+A6
How do you "discover" that user input can be used for injection attacks?
◧◩◪
3. upward+Ak[view] [source] 2023-11-19 10:54:44
>>hhjink+Nj
Good question. We were the first team to demonstrate that this type of vulnerability exists in LLMs. We then made an immediate responsible disclosure to OpenAI, which is confirmed as the first disclosure of its kind by OWASP:

https://github.com/OWASP/www-project-top-10-for-large-langua...

In the citations:

14. Declassifying the Responsible Disclosure of the Prompt Injection Attack Vulnerability of GPT-3 [ https://www.preamble.com/prompt-injection-a-critical-vulnera... ]: Preamble; earliest disclosure of Prompt Injection

◧◩◪◨
4. vasco+mR7[view] [source] 2023-11-21 07:09:10
>>upward+Ak
You "discovered" trying to fool a chatbot? It's one of the first things everyone does, even with old generation chatbots before LLMs.

If so then 4chan had prior art, discovering prompt injections when they made Microsoft's Tay chatbot become a racist on twitter.

[go to top]