ā€œThe LLM generated what was described, not what was needed.ā€

https://blog.katanaquant.com/p/your-llm-doesnt-write-correct-code

Your LLM Doesn't Write Correct Code. It Writes Plausible Code.

One of the simplest tests you can run on a database:

Vagabond Research

@jack

There's one line there that would be valuable in a broader discussion (beyond coding):
"This is not a syntax error. It is a semantic bug:"

Given that LLMs don't even have a semantic component, this seems foreseeable. But outside the specific context of coding I see people struggling to formulate the point cogently.

This would be relevant to contexts like technical writing, etc., but I don't think the vocabulary exists for expressing it neatly in such cases.

@glc @jack

When people start out with LLMs, they ask for code.

But after a while, the users ask for spec.

@jack

This is gold:

In the 1980 Turing Award lecture Tony Hoare said: ā€œThere are two ways of constructing a software design: one way is to make it so simple that there are obviously no deficiencies, and the other is to make it so complicated that there are no obvious deficiencies.ā€

@jack Unfortunately, I've had the honor working with senior developers who operate in similar way. I mean, they produce code much faster than they think (or that anyone else can read/review). Creating huge "working" unmaintainable code bases that reinvent every possible wheel. Sales and management are super happy. Then someone else have to spend years cleaning up their mess.

They are senior and super fast so everyone has to adapt. After all, these devs produce lot of code and must be pro...

@jack I recently spoke to one of these developers. He claimed that his greatest use of coding LLMs is to reduce the size and complexity of the code he writes. He simply can't stop himself writing insane amounts of code. I'm quite the opposite. An interesting anecdote I believe. People are different.
@jack My current favorite is ā€œGenerate a list of this kind of organizations (there are n of them) and their CEOsā€
ā€œI created a list [well done], but I couldn’t find all CEOs, so I guessed the rest based on the names I found!ā€
The result looks plausible, of course, so the LLM did its job :-)
@chris They are not grounded in world models, so poor performance on tasks involving facts is not very surprising šŸ¤·ā€ā™‚ļø