Andreas Wagner

@anwagnerdreas@hcommons.social
1,078 Followers
1.4K Following
5.2K Posts

I am #DigitalHumanities Coordinator at Max Planck Institute for #LegalHistory and #LegalTheory (#mpilhlt) #Frankfurt

Also collaborator of salamanca.school project of #adwl #Mainz

While I mostly toot about work, I do have hobbies...
#Capoeira #Stratocaster

#LawFedi #Histodons
#NLP #TEIXML #Golang #Python #Elm #XQuery

If you're reading this on bsky, follow @ap.brid.gy so i can see your replies.

orcidhttps://orcid.org/0000-0003-1835-1653
hcommonshttps://hcommons.org/members/anwagnerdreas/
pronounshe/him
My overall reading: if this is the law, we can feel confident that 1) libraries that physically possess copyrighted books 2) can let researchers train models on them 3) without a lot of fuss about exactly what % of any book a model would be able to reproduce 4) and even share those models! +
Oh, and if you want to consult the original, see below. Glad this did not settle out of court. That would have really screwed over academic researchers and libraries without big legal staffs. bsky.app/profile/eriq...

RE: https://bsky.app/profile/did:plc:i6w2uuipovifyhfu62viemv7/post/3lse3ncj4qk2q
seems to imply that the liability in that case would rest with Hugging Face or whatever service runs the model / uses it to infringe. Having created a model that *could* infringe is not actionable here. Finally, not a practical point, but this analogy will piss ppl off and I'm glad.
Central practical detail. It matters a lot, in this ruling, that Anthropic did not serve their model in a way that produced infringing copies. Having the potential to do it is not illegal. The gray area here for libraries will be "What if a model is mounted on Hugging Face?" But Alsup's ruling +
I'm reading Bartz v Anthropic, of course, from a very practical, interested perspective. ("What can libraries let me do, and how exactly can we stay out of jail?") The first thing I notice: I didn't realize Anthropic had rolled their own GBooks. Doesn't sound prohibitively $$$. 1/3

However, at the point I was asking I actually did not want to address these more formal considerations yet. I was trying to get a more comprehensive grasp of the subject matter and its different dimensions. To this, the conversation actually did not contribute a lot.

#AcademicChatter #Chatbots

3/3

Helpful: merely having a sounding board was helpful. It suggested some aspects that I then excluded more consciously from my draft and brought up ways of structuring the text some of which I then adopted.

However, I am not sure how much of the "sound" of the "sounding board" I am projecting onto the bot or the dialogue. I think it was basically my own reflections of the initial text and its stakes that really improved the document. The only input that must objectively be attributed to the bot was the pointing out of aspects and structure of the document typical for the genre.

#AcademicChatter #Chatbots

2/3

Today I figured out why I often feel uneasy with chatbot responses. They are helpful and disappointing at the same time: About some draft I asked: "did I forget an important aspect?" and when I pointed out that I considered the bot's subsequent response to be indeed covered by what I had written initially, the bot (Claude 4 Sonnet) just gave in: "Yes, now I admit that you're 100% correct etc.".

Disappointing: I think I was hoping for some friction as I would expect from a human interlocutor who would probably have had an intuition that my coverage was only partial, or bound to some unarticulated presupposition. In my experience, such a kind of friction is almost impossible to find in a conversation with a bot.

#AcademicChatter #Chatbots

1/x

@zackbatist Right, I agree, these are important considerations. And it’s necessary to balance that with additional sources. (My second source are usually hand-crafted thematic library indices.) But if it helps identify relevant English language journal articles from the last 15 years that I would have missed otherwise, I feel that’s already useful.
@felwert Very poor quality for monographs, chapters in edited volumes and grey literature, and for anything not in english or from before the past ~15 years
×

I like this take by @kentbeck on how AI-assisted programming changes the balance of which skills are most important

From this interview with @gergelyorosz https://newsletter.pragmaticengineer.com/p/tdd-ai-agents-and-coding-with-kent

@simon @kentbeck @gergelyorosz There’s certainly something to it, but a real problem with AI assisted programming today is related to the old adage that if you write code as cleverly as you are capable of, you won’t be clever enough to debug it. AI-assisted programming hugely amplifies that challenge. Perhaps it goes away as the tools get better, but today it’s a problem.
@simon @kentbeck @gergelyorosz Do they discuss how people are going to acquire those 10% in the future (I assume those are the advanced seniority 10 percents?)

@helge @kentbeck @gergelyorosz that's one of the most interesting questions right now, I think

We clearly need to reimagine aspects of how we train new software engineers

@simon @helge @kentbeck @gergelyorosz my friend mentioned the other day we should also think about "training the AI". Should I publish books that are written for AI and not books written for people?
@chris @simon @kentbeck @gergelyorosz If you get paid for that, sure 🙂 I still think AI is a huge issue because of the "stealing" part (doesn't matter what you call it, but I think we all understand the issue and that it leads to content not being published anymore publicly).
It's essentially the other side of the seniority coin, if everyone just synthesises code, there is no new input for number 5.
@helge @chris @simon @kentbeck @gergelyorosz I'm sure that StackOverflow is a huge input for programming information, and I know I'm not alone in dramatically reducing my time there. After writing over 5700 answers over 15 years, I haven't answered anything in 2025. I do expect this to become an existential problem for the models across a lot of fields.
@helge @chris @simon @kentbeck @gergelyorosz But to Beck's point, I think to a first order, you should think of coding assistants like a higher-level programming language, not a complete reinvention of programming. I find that most of the usual skills still apply, even when they're running at their best. And to "how will junior devs learn the low level skills I know," I'd say the same way most devs learn assembly language. They don't. And it's mostly fine.
@helge @chris @simon @kentbeck @gergelyorosz My much bigger concern, having used coding assistants quite a bit now, is that they have the ability to really trash a code base really fast when they get confused, which is often. And I expect that this will be a major problem. I really like using them at the very start, but they tend to go off the rails pretty often, and I have to take control back. I expect that will give plenty of experience to junior devs.

@cocoaphony @chris @simon @kentbeck @gergelyorosz I actually think it is mostly the same issue, just at a much, much lower speed. Someone who knows assembly usually writes better code (not because she is older, but has a solid understanding of how things work internally, though admittedly that may be counter today).
But AI brings that issue (not understanding what happens) from an acceptable level to 100x.

Either way, I think it can be helpful to senior devs because they can rate the output.

@helge @chris @simon @kentbeck @gergelyorosz Of course when higher level languages were first developed, senior engineers did not feel that the issue was an acceptable level :D

I'm seeing some junior devs get in way over their heads following AI advice. (I'm kind of developing a stock lecture about it…) I'm also seeing junior-to-intermediate devs use AI to explore and learn deep things they wouldn't have dared before. I'm seeing the dig into details that before they'd have left as unknowable.

@helge @chris @simon @kentbeck @gergelyorosz (And yes, AI hallucinates. And also, when you research things on the internet, the internet hallucinates. And when you study things in books, they also sometimes are just flat wrong. There are definitely differences, but it is not a fundamental break with the past.

@cocoaphony @helge @chris @simon @kentbeck @gergelyorosz the difference (to my mind) is that if you have a solid foundation in “how things actually work”, you can design an experiment to figure out what is really happening so you don’t get misled. Absent that foundation … all you can do is hope.

That doesn’t have to mean “writing assembly,” but “knowing that there’s an actual spec and knowing how to read one” is an _invaluable_ skill that’s severely diluted by vibe coding. Even the idea of having a spec is diluted.

@steve @helge @chris @simon @kentbeck @gergelyorosz Also true. Though time and again I discover that how I think things actually work, based on my decades of foundations, is...huh, not actually how things work (which I learn when someone like Steve explains it to me... :D)

Certainly, learning foundations will always be valuable. But I remember my mentor in 1996 fussing that I shouldn't waste so much time digging into details. I think they were wrong. But it's not like *everyone* used to care.

@cocoaphony @helge @chris @simon @kentbeck @gergelyorosz The thing I worry about losing (and which we had already lost to some extent pre-wet-hot-AI-summer) isn't any specific skill or knowledge, but rather the common agreement that computer systems are deterministic things and you can figure out how anything works (or why it doesn't work) by a combination of deliberately reading a specification and conducting experiments to validate it.
@steve @helge @chris @simon @kentbeck @gergelyorosz But I am with you about "vibe coding." I expect that's a fairly short-lived thing. When I see people who are really successful with it, it turns out there was often a *lot* of planning that went into that "vibe." :D
@cocoaphony @steve @chris @simon @kentbeck @gergelyorosz I wouldn't underestimate that. It isn't that different to "regular" users hacking up Excel macros or say Shortcut scripts (which usually s*** from a dev pov, but they do the job). It is an enabler, and as long as the users can properly rate the output, that is kinda ok (seniority again).

@helge @steve @chris @simon @kentbeck @gergelyorosz Agreed that coding assistants are definitely lowering the bar (in a good way) for non-devs to build their own tools. I've seen quite a lot of that, and it's a great thing.

I think some folks are scaling that incorrectly to "AI, build and deploy a replacement for Netflix, I'll come back in an hour." This is very related to how folks confuse prototypes with "almost ready to ship."

@cocoaphony @chris @simon @kentbeck @gergelyorosz Of course that always was a thing. The level w/ AI is *way* higher, like 100x (or 1000x according to the article) vs like 2x or maybe 10x. Entirely different scale and problem domain.
Suggesting that this is just the regular "new abstraction" complaint is distracting from the broadness of the issue, IMO.

@simon @kentbeck @gergelyorosz I don't write code. I am more of an _idea_ person.

Good luck with that.

@lemgandi @simon @kentbeck @gergelyorosz If what you’re saying is that there’s not a lot of value in being an “ideas person”, i.e. someone who designs and architects code based projects, then I can see that too. Because just as LLMs have wowed people into thinking that the basic skills of programming are now pretty much worthless (as in the original quote), I have absolutely no doubt that in the next couple of years it will be possible to get good enough ideas-type work out of agentic AI as well. People will share their rule files and whatever other crap, and alongside enhanced context awareness it’ll be the same thing… so then where’s your value as a human programmer? Enjoy “talking” to an IDE all day long and nibbling at the edges of design decisions and code alike?

@jimgar @simon @kentbeck @gergelyorosz Ideas are very cheap. How 'bout a website where you can design a door? Or one that helps you find the best food for your dog? Or a site to help you fix your shoes? Or a phone app that will automatically order your lunch from the nearest fast food restaurant?

Implementation is hard. Writing a real application that can actually withstand malice is hard. Understanding the real world is hard.

@simon @kentbeck @gergelyorosz You are all programmers in this thread. I am not, but see how the quote of Simon’s original post is mostly applicable to general white collar work too. Lots of ”what is done” is just repeating patterns that have been robustly done before (thus - can be tackled/helped by LLMs), while a very small percentage of work (varies with personality and role) is applying judgement in context that is unique. (1/2)

That judgement needs to be practised while getting to grips with the fundamentals.

To every junior, the world is black and white (digital), while over most folks lifetimes shades of grey appear (analogue). (2/2)

@simon
I think that's extremely optimistic because the most consuming part of any creative development is specification (ie a language of description to define the operation), in ever increasing precision until the point where the machine does what you want.

You can't handwave. Anything an LLM generates has to be verified by someone or something against...what? -> against an understanding?

An LLM can help you find some steps quicker but it can't do that *work*.

@kentbeck @gergelyorosz

@happyborg it is intended to be optimistic. Sounds like you mean “unrealistic”. How?
@simon @kentbeck @gergelyorosz really nice clarification on what he meant at that time 🙏