Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You didn't respond to this

> Let's assume that in 10 years, LLMs are at the point where you cannot distinguish between a well written book by an author and one generated by entirely by an AI. Suppose we have two authors, one that is long dead and whos works are public domain and a young one that is just starting. An LLM trained on the author whos work is in the public domain can generate books that is just like the original works. But what if the young author writes in a similar way, is that now legal or illegal to generate the same content? It's impossible to know if the young authors work have been used for training.

which is central to my reasoning. You could go the same way as software patents, but that is not preferable in any way.

---

>> I am certain that trying to ban LLMs, or dictating what and how is not the answer.

> I wouldn't ban LLMs because of copyright issues, though I would let authors choose whether their IP can be used for training or not.

> Why not? Just say that using for training is considered derivative work, and that's it. Now copyright owners just have to update their license to allow for training if they want to, and that's solved. Of course, Big Tech makes less money from that scenario.

Big Tech can train on everything that is "legal" and malicious actors can finetune with a specific authors works and then generate books. You will not be able to detect that and the malicious actor can claim to have written it themself. Then we're back to the starting point.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: