As I understand it, the state of copyright for LLM-generated output is:
So accepting LLM-generated code seems very high risk. Especially for something like ext4, where it’s likely that there is exactly one implementation in the training set.
EDIT: I don’t expect the second to be settled conclusively any time soon, because there’s a lot of subjectivity in existing human-only derived-work cases.
I agree that the second won't have a court ruling anytime soon, but the fact that some AI companies are settling instead of going to court is a strong signal. If they truly thought they'd win in court, they wouldn't settle for $1.5b. (yes there is risk reward to consider when analyzing this, but that just, imo, shows that the legal standing is even weaker, that if they lost they'd likely be on the hook for even larger penalties.)