Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

That sounds like something I could get an LLM to do. And then of course I can do it iteratively until all the code has been laundered. Maybe that's how Microsoft can justify training on all the GitHub data.


IANAL, but my understanding of copyright law jurisprudence is that using an LLM to automate the process is going to substantially increase the likelihood that you will be found to be infringing.


I think, but I'm not sure, that they mean to write the tests, and then they'd be able to fix the implementation blindly?


Except the detail about what licenses were used for the learning model.

It remains to be tested in court.


It’s entirely possible that the model and all of its outputs will be determined to be derivative works of the training inputs. If that happens then, oh boy, not good things for anyone using it, I’m sure.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: