cool guess we need to start the alpine "are upstream projects with LLM contributions legal for us to ingest" conversation
my personal guess is no, if you were wondering.

@ariadne

Well the best approach would probably be to say "no for now". But following recent court rulings (I think it was in the US), the answer could change to a "yes" as it appears that they see that any code generated with AI is basically public domain as copyright is for human generated work only.

@agowa338 I would like to see a respected organization like Conservancy give some guidance here
@ariadne @agowa338 AI can't copyright-wash the code it spits out. The chatbot will closely crib an existing example from its training if the thing you ask it for isn't generic. That's the copyright threat model here.

@davidgerard @ariadne

Well the latest court ruling I saw a few days ago was that when they used AI generated code in a codebase with a specific license that any code that was AI generated was public domain and that license didn't apply.

But you know, it is still quite young and both courts and politicians take time to find "a stable and predictable ruling", so...

With, say, art, if I take a photo of a painting, then there are two separate copyrights to deal with - the original copyright in the painting and the additional copyright in the photo. I own the 2nd, but not the 1st - if you want to reproduce my photo, you need two licences to do so, one from me, one from the original artist.
LLM output may not be copyright, but it shouldn't remove the copyright from everything the LLM ingested and regurgitated into its output.
@agowa338 @davidgerard @ariadne

@BenAveling @davidgerard @ariadne

That wasn't part of the court ruling. Everything is still heavily in flux. So the only thing that is clear right now is that nothing is clear...

@agowa338 @davidgerard I would like to see an actual organization such as Conservancy provide any guidance on this topic.
@ariadne @agowa338 it would be nice! i fear the actual answer will be "facts and circumstances"

@ariadne @agowa338 without legal precedent everything is just guess work. Until we get a SCO vs IBM sized case nothing's for sure and the LLM bubble blowers are going to do whatever they can to prevent such a case from happening any time soon.

Their strategy is when it is in everything it needs to be declared legal.

I hate this timeline.

@TheOneDoc @ariadne

Meta made it "fair use" to mine the entire web for data and train you AI on. But that was probably not the case you were thinking off...

@ariadne How are you going to deal with critical packages that allow LLM contributions, like Linux?
wdym by "ingest" here ?
@fiore @ariadne use code thereof i would believe
so not "include in official repos" ?

CC: @[email protected]
@fiore @ariadne probably

i think the logic would be that llm code cannot be assigned legal authorship and therefore any license permitting the redistribution thereof would be void
@fiore @coolbean that is the concern, yes. and also the possibility of misattributed code recreated by the LLM.
@fiore @coolbean in other words, I am not certain it is legal for us to redistribute harfbuzz 13.0, as an example.
@ariadne @fiore harfbuzz too? god im going back to bitmap fonts fuck this

as in the projects packaged in the repo?

@ariadne

@ariadne I hope you don't mind me paraphrasing this to "should Alpine take the legal responsibility to upstream things none took any legal responsibility until now?", I know it is not what you said but this is how I have experienced this process, and a default no is the only logical answer atm.

given though that we now have repos that hide AI contributions as well, it is a clear indicator that even outside the Alpine scope this is not a code contribution subject, it is a broader liability one, and a need to discuss what will happen if even involuntarily a maintainer ends up upstreaming something "legally/security/community toxic" is proper.

Most ( tech friendly ) legal people I spoke with just end up with "just avoid clear trademark infringement and set up an integration framework to prove to a court you at least tried if shit hits the fan", which I personally take it as as hard no for prod readiness.

Just to clarify, I am not against AI used as productivity tool in general ( same way as I am not for or against using an IDE to write code ), but I am definitely not ok setting up horizontal ( or in fact any ) rules upon castles made of sand... especially where it is not fit to do so and will introduce risks on the principles of a project.

Maybe after we, as society, go through some law suits and have established a better foundation on this subject.

@ariadne What about Linux and LLVM? I think those are far more serious problems than harfbuzz.