zlacker

[return to "GitHub accused of varying Copilot output to avoid copyright allegations"]
1. Shamel+ba[view] [source] 2023-06-10 14:56:50
>>belter+(OP)
Eh, their argument is simply that they tuned temperature settings to encourage the model to output slight variations on memorized data. But this is kind of just one of many things you do with a language model and certainly doesn’t imply intent to avoid copyright allegations.

Just implies they tuned it for user experience.

I was expecting there to be some discovery around them deliberately fine tuning their model to output modifications if and only if the code had a certain license.

◧◩
2. Brian_+Ud[view] [source] 2023-06-10 15:20:46
>>Shamel+ba
Why else bother with such an input? Are randomizations more likely to be correct or more useful?
◧◩◪
3. seanhu+vJ[view] [source] 2023-06-10 18:17:34
>>Brian_+Ud
Generally the reason behind adding randomness to machine learning is avoiding "local minima" in the search space of the optimization function(s) used for training the model. If your training produces a very smooth descent to an optimum it can lead to the model converging on a solution that is not globally the best. Adding some randomness helps to avoid this.

Specifically for GPT models, the temperature parameter is used to get outputs wihch are a bit more "creative" and less deterministic. https://help.promptitude.io/en/ai-providers/gpt-temperature

[go to top]