I was thinking about "AI-assisted" coding, legal risk, and what it would mean to allow only "trivial" assistance (like formatting or "find the bug" interactive sessions) while disallowing "significant" contributions (like whole subroutines or larger blocks of code). I personally don't think it's possible to set aside the moral component of these models (stolen training data; excessive land, power, and water use and general "bad neighbor" habits; monopsony power in computer component acquisition) and I remain opposed to their use for that reason. But I think we have to assume that people who are already willing to overlook the moral component may also be willing to lie about provenance in order to land a contribution to a project.

Further, there are several large open questions about copyright in relation to agents. Is their training data truly unencumbered by copyright and licensing restrictions? Is their output copyrightable? Does even a "trivial" contribution by an agent invalidate the copyright of a larger code contribution that contains it? Is the user of a coding agent, or the user of the software generated by that agent, indemnified if the resulting code is later found to infringe a copyright or violate a license? Can an agent truly be said to output a "clean room implementation" of something when there is a non-zero chance that its training data contained the thing being reimplemented, and there is no way to verify that?

So, in general I'm against coding agents on moral grounds, and I'm also against them on legal grounds because I think any risk at all is too much risk. But on the other hand I'm intrigued by the question of "trivial" contributions, and I suspect that even projects that don't allow assistance from AI coding agents may have unwittingly accepted code that contained such "trivial" contributions. My questions are:

1. Is it possible for an AI-assisted code contribution to be "trivial" enough that it presents no legal risk, either now or in the future?
2. If so, how would you go about determining what's "trivial" and what's "significant?"
3. How could a contributor not just self-certify, but present verifiable evidence that a code contribution was legal and non-infringing and that any contribution from an agent met the "trivial" standard?
4. How could a company or open source project protect itself against a dishonest or bad faith actor who contributes code that later is found to infringe on a copyright or violate a license?
5. Who's going to pay for the damage if the worst case scenario comes to pass?

I don't have answers, but I suspect that the question of what constitutes a "trivial" contribution is going to matter a lot in the future.
Oooo, fascinating article (via @aparrish) that dovetails with my questions about calculating the risk of AI coding agents, but starting from a report that said "Another looming problem is that large insurers have become wary of underwriting policies that cover companies against AI risk." I guess I'm not the only one thinking in this way. https://freakonometrics.hypotheses.org/89367
If No One Pays for Proof, Everyone Will Pay for the Loss

This post was initially written in French, Si personne ne paie pour la preuve, tout le monde paiera pour le sinistre Let’s start with a truism. In ordinary life, just as in economic life, we have to make decisions without ever knowing everything. Every decision involves some uncertainty, and therefore some risk. Some risks are … Continue reading If No One Pays for Proof, Everyone Will Pay for the Loss →

Freakonometrics

@fedward @aparrish IMO the only way you can go all-in on AI is if you've used AI, been awestruck by its output, and never actually verified its output.

Insurance companies don't strike me as being run by the type of people who miss that last step. Unfortunately, so many people are so overwhelmed by step two they never proceed to step three. (Or they spot check it once when it happens to be right and then assume it always is.)

@fedward @aparrish meanwhile my corp masters are ALL IN ....... scary all in...CEO has been main lining the Kool-Aid

@fedward I think you are more thoughtful than upper management in my day job......they are 100% in and forcing more AI and AI enabled automation down the line.....and they want it all done NOW to should "productivity gains" and justify the ridiculous $$$$ being spent.

the legal question was probably brushed aside b/c our lawyer are BIGGER than anyone who might sue. BUT also all the executives who chose this path will have moved on by the time there are any consequences....

Oh and my use of AI coding agents always finds them wanting when you actually know what you are doing.

@fedward I’ve been a bit depressed about the future of open source. It’s feeling like there won’t be a way to prevent clones so business-backing will dry up and anyone looking to maintain rights is going to struggle with both asserting their license and preventing clones.
@acdha what even are copyrights and licenses anyway