when you buy a book, you are not acceding to a license to only ever read it with human eyes, forbearing to memorize it, never to quote it, never to be inspired by it.
> Interestingly, Llama 1 65B, a similar-sized model released in February 2023, had memorized only 4.4 percent of Harry Potter and the Sorcerer's Stone. This suggests that despite the potential legal liability, Meta did not do much to prevent memorization as it trained Llama 3. At least for this book, the problem got much worse between Llama 1 and Llama 3.
> Harry Potter and the Sorcerer's Stone was one of dozens of books tested by the researchers. They found that Llama 3.1 70B was far more likely to reproduce popular books—such as The Hobbit and George Orwell’s 1984—than obscure ones. And for most books, Llama 3.1 70B memorized more than any of the other models.
Memorising isn't wrong but when machines memorise at scale and the people behind the original work get nothing, it raises big ethical questions.
The law hasn't caught up.
I also play the guitar, and it took me 10 years to learn 30 or 40 songs. So I don't see how anyone can learn 7 million songs in a couple of minutes.
Most AI seems much better at reproducing a semi-identical copies of an original work than existing video/audio encoders.