zlacker

[return to "GitHub Copilot available for JetBrains and Neovim"]
1. ipnon+V5[view] [source] 2021-10-27 18:11:22
>>orph+(OP)
Can any users give their opinion on how it's helping their productivity? What problems are they finding, if any?
◧◩
2. etaioi+E9[view] [source] 2021-10-27 18:26:36
>>ipnon+V5
I only played around with it on OpenAI but it's the same model as far as I know. It's pretty good at regurgitating algorithms it's seen before. It's not good at all at coming up with new algorithms.

It's very good at translating between programming languages, including pseudocode.

It can write a lot more valid code much quicker than any human, and in a whole slew of languages.

I haven't had the urge to use it much after playing around with it constantly for a few days, but it was pretty mind-blowing.

◧◩◪
3. yeptha+Ja[view] [source] 2021-10-27 18:31:19
>>etaioi+E9
Your response makes me wonder if poisoning the well is possible by submitting code to Github with multiple languages and coding styles. A single file with a function signature written in Javascript and the body written in Python + Ruby. Enough code would surely break the AI model behind it. Unless Copilot has some sort of ingestion validation which wouldn’t surprise.
◧◩◪◨
4. ctoth+EG[view] [source] 2021-10-27 21:07:26
>>yeptha+Ja
I don't know if this is true, but I would assume that the tokenizers they used for Codex use actual language parsers which would drop invalid files like this and make this attack infeasible.

When I was playing around a couple years ago with the Fastai courses in language modeling I used the Python tokenize module to feed my model, and with excellent parser libraries like Lark[0] out there it wouldn't take that long to build real quality parsers.

Of course I could be totally wrong and they might just be dumping pure text in, shutter.

[0]: https://github.com/lark-parser/lark

[go to top]