zlacker

[parent] [thread] 4 comments
1. jefftk+(OP)[view] [source] 2022-10-17 10:58:08
In case anyone interprets this literally: if copilot regurgitates literal code it was trained on that doesn't actually give you an unencumbered version.
replies(1): >>vasco+12
2. vasco+12[view] [source] 2022-10-17 11:15:39
>>jefftk+(OP)
So how long till new software licenses that prohibit any use of code for model training purposes? I'd be willing to bet there's a significant group of people that won't be happy either its literal or not, the fact that it was used in the training might be enough.
replies(1): >>jefftk+15
◧◩
3. jefftk+15[view] [source] [discussion] 2022-10-17 11:38:56
>>vasco+12
The claim that most people training models make is that what they are doing is sufficiently transformative that it counts as fair use, and doesn't require a license. That means putting something in a software license that prohibits model training wouldn't do anything.

In this case, what the model is doing is clearly (to me as an non-lawyer) not transformative enough to count as fair use, but it's possible that the co-pilot folks will be able to fix this kind of thing with better output filtering.

replies(1): >>Feepin+O7
◧◩◪
4. Feepin+O7[view] [source] [discussion] 2022-10-17 12:02:01
>>jefftk+15
Prohibiting training would not affect the produced source, but it would make the training itself illegal.
replies(1): >>jefftk+O9
◧◩◪◨
5. jefftk+O9[view] [source] [discussion] 2022-10-17 12:17:24
>>Feepin+O7
Probably not? Models are trained on restrictively licensed things all the time, such as images that are still in copyright. This is generally believed to be fair use, though I think this has not been tested in court?
[go to top]