I think this, a discussion of the parallels between "AI" and "crypto", is a good take. I want to dig into the bit on "AI" being different because it has practical use.

"AI" is a marketing term. There's the stuff that was mainly called "ML" up until 2021 or so, which definitely has practical uses. E.g., if you're running a social network and need to help humans find the toxic stuff, ML can help.

But in the last few years there's a wave of hype mainly around the large language models, LLMs, and the large text-to-image models. So things like ChatGPT and DALL-E. It's really not clear to me those have much more practical use than crypto. Certainly not over their costs. 1/

https://sfba.social/@[email protected]ial/111754701923644674

Jesse Baer đŸ”„ (@[email protected])

People like to say that "AI" is different from crypto in that there are actual useful applications, and that's true. But the vast majority of people you're expecting to come up with those applications are the same people who were just trying to build products on the blockchain.

Mastodon

Just to be sure I'm not being unfair, I searched for writeups of LLM uses. Here's a representative example of the genre: https://www.techopedia.com/12-practical-large-language-model-llm-applications

They mention 12 use cases. Some are done better with simpler, cheaper models or non-ML techniques (1, 4, 6, 8, 12). Some are wildly speculative (2, 9, 10).

But that leaves 4 items that I want look at carefully: content creation, customer support, sales automation, and writing code. Those are at least superficially plausible places where LLMs and large image models could have practical uses.

2/

First, though an important philosophical point: LLMs are fancy autocomplete. You give them a set of words, they'll predict the next word based on the enormous corpuses they've been trained on. This can give them the appearance of sentience. People will talk as if they "understand" things. They don't. It's the million-monkeys-with-typewriters thing, but the monkeys have seen enough English text that the next word is statistically consistent with the previous words.

Humans are subject to to pareidolia, and we really like to anthropomorphize things. It's not just thunder, it's a guy with a name and a look and a personality and a whole family. It's not just a bit of winter where we celebrate with our dearest; it's a fat guy in a red suit with specific facial hair. So although the text can feel human, we'll have to work hard to think of ChatGPT as a bit of unfeeling machinery, not our plastic pal who's fun to be with.

3/

Ok, so first, content creation. That seems positive, right? Wrong! The best way I've seen of explaining this: "Why should I take the time to read something nobody took the time to write?"

I think this one is a huge net societal negative. The people out there who want instant "content" are almost entirely not readers. They're people who want something to run ads against. They're people who want the credit for writing without doing the work. They're people who want to sell you something without understanding the something or whether or not it might be good for you. In short, they're people with various levels of contempt for their readers.

4/

As a writer, I think the real value of writing is the thinking and care that goes into it. Even for purely factual material, writing involves a careful search for truth.

LLMs, though, don't have any concept of "true". They can't. What they have is the digested correspondence between words in Wikipedia and Reddit and a zillion other sources of text. Truth can be represented in text, but it lies outside of it.

In philosopher Harry Frankfurt's "On Bullshit", he defines it as "speech intended to persuade without regard for truth": https://en.wikipedia.org/wiki/On_Bullshit

Marketing content generated by LLMs is clearly bullshit. But I'd argue that by imitating human forms of writing, *everything* produced by an LLM is bullshit. (Which would make the enormous "AI" hype cycle bullshit about bullshit, a truly American accomplishment.)

5/

On Bullshit - Wikipedia

Let's turn to the second plausible use case: customer support. People often like to talk to other people to resolve problems. What if we can automate the *feeling* of talking to a person, but with no actual people involved?

There are a bunch of things going on here. At least here, there's a real user need. But to what extent is it a real user solution?

It's plausible to me here that this will be a plausible first-query solution once you've built up a good base of Q&A examples for it, doing a bit of textual generalization. At least as long as what you're doing is pretty standard. But remember that we're using a bullshit engine here, so what happens when the statistically plausible text isn't correct or useful?

6/

A good example here is the car dealership that tried using a ChatGPT bot for customer service. It quickly agreed to a "legally binding" offer to sell cars for $1: https://venturebeat.com/ai/a-chevy-for-1-car-dealer-chatbots-show-perils-of-ai-for-customer-service/

Would any human agent do this? No. Because humans understand things. This is a toy example, but if GPT-ish things don't work for very basic cases, how much can we rely on them for important cases?

7/

A Chevy for $1? Car dealer chatbots show perils of AI for customer service

Incidents at car dealers highlight the responsibility of ensuring target chatbot deployment and safety compliance. 

VentureBeat

Moreover, I think there's a deeper problem here. Anybody who's ever tried to use large-company customer service (and I'm thinking of Amazon here) might say, "well a chatbot couldn't do much worse; it's not like talking to a person right now."

There's a really handy concept called "failure demand". Some load on a system is "value demand". E.g., I walk up to the counter in the local store and say, "I'd like a quart of milk." But if have to come back later because they're out of milk or the milk was spoiled, that's "failure demand".

For digital products, I'd argue most customer service demand is failure demand. So generally we shouldn't be putting ChatGPT in to tell people how to solve their problem, we should giving them good experiences in the first place. Chat's a bandaid.

And bad customer service generates even more failure demand. So ChatGPT might lower costs, but it won't solve the root problems or on net improve things for customers.

8/

Ok, what's next? Right, "sales automation". This is a bit of a mix of the first two cases. "Sales" is a little bit customer support, and a lot of manipulation to get people to buy things. Setting aside the possible minor customer-support improvements, I think we're mostly back at content generation with contempt for the reader.

One thing people have a hard time grasping is how much our economy spends on manipulating people to get money out of them. Advertising alone is hundreds of billions. Sales is surely at least that much. And then there's all the time lost to fending off the manipulation, plus the money lost when people fall for it. On the order of magnitude of the defense budget or all K-12 spending for sure.

So a bullshit machine might be a good fit, but is this a hole we really need to dig deeper?

9/

And lastly, we have writing code.

I've been writing code since I was 12. I started writing code for money at 18. And my dad started making his living as a programmer in the 1960s. And this all has a familiar ring to me.

As a young teen, so maybe 1982, I went to my first tech conference. Walking the exhibit floor, I found somebody selling a system that promised to eliminate expensive programmers by letting the business people just write in plain English.

It's an old dream, and one I've seen come up many times. Visual programming systems. Code wizards. Model-Driven Architecture. Probably many waves I've missed.

10/

Of course, we also have the dream appearing regularly in science fiction. HAL 9000 is from 1968, and was set in 2001. You can just tell the computer what to do, and off it goes, understanding your needs.

I should say that this is a lovely dream. And it can be a great inspiration as we get machines and computers to do the drudgery. But as with all powerful dreams, we have to be careful to be realistic.

11/

When people expect "AI" to do the coding, I think they're not paying much attention to what the hard parts of software development really are.

The computer I'm typing on is at least a million times better than where I started. A million times faster, a million times bigger. So much of the work I did then has been automated. But the job hasn't gotten easier, it's just a different kind of hard. Instead of me hacking away solo in my dad's basement, I now am trying to collaboratively build lasting, coherent intellectual works directly with my team and in concert with thousands of other people via libraries, services, and the like.

12/

If I need to knock out a little boilerplate in a language I don't understand? Sure, I will happily use a bullshit generator for a chickenshit job.

But if I need to build anything that lasts, then I'm going to have to do the work myself, because the major part of the work is in *understanding* the situation (the users, their needs, the team, the history, the code), and carefully improving it, not in just typing shit out.

Is it possible that tools can accelerate this? Sure. I've been using an IDE as brain augmentation for 20+ years. I think it would be amazing if I could automate myself out of a job. But at best here I think we'll see some improved autocomplete, and at worst I think we'll see, as we have before, people generating absolute reams of garbage code that other people then have to maintain and clean up.

13/

So in sum, I think the currently expected positive uses of "AI" are either things ML was already doing or possible minor improvements.

But a bunch of those improvements are on top of systems that should be rebuilt to be radically better, not just slapping another layer of shiny wallpaper on top. And a lot of what people are going to do with the ability to generate "content" is a net societal negative. Manipulation, disinformation, propaganda, and a general sea of bullshit that real humans will waste time reading.

Alas, as we've seen with the 10-year arc of "crypto", we appear to be willing to waste approximately infinite time and money as long as a) there's a hard-to-refute technoutopian gloss on top, and b) somebody somewhere is pocketing some of that money.

14/

And in closing, let me recommend Zac Bissonnette's 2015 book, "The Great Beanie Baby Bubble: Mass Delusion and the Dark Side of Cute". It does a really good job laying out how bubbles work. It's a fun read and it will make you more able to spot the pieces of a hype cycle as it gets into motion.

15/15 (for now, at least)

@williampietri adding to the list!
@damoncjones It's so good. I read it on a lark, but it's surprising to me how fresh it seemed given the "crypto" and "AI" bubbles.
@williampietri same vein, I recommend Attack of the 50 foot block chain. Came out a few years ago, still relevant.
@damoncjones Totally agreed! I'm a big fan of @davidgerard's blog, too!
@williampietri @davidgerard thank you for providing his Mastodon link, I totally forgot to look him up. Great writing.
@williampietri fantastic thread, thank you for putting in the time. 

@williampietri A very nice thread, I had to take notes to keep everything straight on the 12 uses.

I guess I have two reactions:

1. I loved your use of "contempt for readers". I think that is very well put on "content generation", and I haven't seen put quite that way before- it seems right.

2. While I agree with the characterization of LLMs, I think the tools we are consuming, be they GitHub-Copilot or ChatGPT, etc, are growing as software systems beyond the LLM which sits at the core of them. It could be this is all cruft which topples the value of the product, or it could be that they'll nail key quality or safety considerations and reduce error/hallucination failures to an acceptable level (whatever that means),

Anyhoo, while the "spicy autocomplete" take is one I align with, I can't help but wonder if the software systems will eventually transcend those limitations of the underlying LLM.

@jjlupa Yeah, it's plausible to me that something like ChatGPT becomes the "mouth" of a richer system eventually. Sort of like the head PR person for a big company, its only job to be glib. But it's also plausible to me that it's an evolutionary dead end, too tangled and opaque to be a good foundation.
@williampietri i find it interesting that the word “bubble” is used for something that is so obviously a deliberate scam (like BBs). but maybe that’s the case for most bubbles, just less transparently so if you don’t have insight in the specific area the bubble is in?

@LambdaDuck One of the things I think is interesting about BBs is that they didn't start out as a deliberate scam. The guy who made them was an obsessive about stuffed animals. He loved them! The initial people who bought them just liked them too. But the whole thing took on a life of its own for all of them. So I think bubble is the right word there.

That definitely happens, though. I think original cryptocurrency enthusiasts were sincere. But the scammers got in early and have been making great money ever since. But even there it can be hard to tell. A lot of Ponzi scams are just people who thought something made sense, got in over their heads, and then did whatever they could to not drown. FTX was like that too; they started out to be a legit business.

@williampietri
Let me summarize.

What makes me đŸ€ź is the hype that makes it necessary, it seems to stick AI onto anything.

When we did many of the things we do today, years ago, we just said the computer did it via an algorithm.

In our company we have been using text classification (ML) for over a decade now. Doing some other very interesting NLP.

Somehow nobody thought about selling it as AI, and beware our data science dept are evil geniuses that will have an AGI next year.

@williampietri
E.g. So AI now can optimize algorithms. Headlines actually have AI write algorithms. Reality is more like AI play games that as side effect result in working implementations. Play the game better, get a better implementation.

What a revolutionary idea, gamification.

Roughly a decade ago, PostgreSQL added SQL execution optimization via a genetic optimizer. Let the best execution plan to survive the arena be used for real. They didn't market it as AI either.

@williampietri By OpenAI standards how AGI is around the corner, current PostgreSQL servers should have already taken control of most satellites around Earth and Skynet is already up and running, it was just so nice not to let us know yet.
@williampietri Not just time and money...real energy and materials. Enough electrical demand to make old coal peakers economical. E-waste output at the rate of a decent-sized country. Aquifers drained for cooling.

@williampietri I think a good example of the bullshit case is people using AI powered tools to "find" exploits and bulk report them to try and earn bug bounties or whatever... meanwhile they're overwhelming small developers and security teams with false positives.

I'm separately curious about code explaining... but again that might work at class or library level but seems much less likely to work at system level where I usually operate.

@ultranurd Yeah, agreed. I'm especially skeptical about it explaining code. Often small differences result in deep changes in meaning. And "meaning" requires coherence, which larger systems often lack. After I spend some time with a new code base, often the "meaning" is in terms of the history or the people involved.

E.g., I was brought in to help with a system that had extensive use of an early version of Docker with custom patches. It was hard to puzzle out the meaning until I found that the person who had put it in not long after did a conference talk on the fabulous new possibilities of Docker, so he'd shoehorned it into a production system and then quickly got a higher-paying job elsewhere. I ripped it all out and replaced it with something much more suitable.

@williampietri @ultranurd What‘s going to be really fun is, what happens as languages get patched? Especially many of the up-and-coming languages still have (minor) semantic changes and extensions of syntax. What happens then? Is an LLM going to borrow incorrect semantics from another language because the syntax looks similar enough?

@williampietri I think the way I think about it, and something I even talk about with me peers in the context of contracts between systems, is syntax versus semantics.

As sparkling autocomplete an LLM on code isn't going to have any meaning around intent/purpose of code beyond how humans label it, just pure what's likely to come next... so boilerplate maybe helps but any business logic beyond that I'm doubtful. And how much of that could be handled just as well by better library encapsulation?

@ultranurd For sure! The way I look at it, any boilerplate is duplication, an opportunity to DRY something up. Copy-paste solutions, whether direct or chatbot-laundered, are a missed opportunity for better abstractions.

@williampietri i guess programming is an area where fancy autocomplete can be slightly useful, as long as you view it as exactly that.

as long as you know what you want and can understand the code it generates, it can be a way to save time and reduce the number of roundtrips to example code and reference documentation (assuming you aren’t missing some important caveats in the docs, but that’s a risk when using ordinary autocomplete too)

@williampietri My most optimistic/hopeful take is an LLM as a pair programmer & rubber ducker. At some point I want to try the free trial of CoPilot in VSCode to see how it works for that. Or maybe the JetBrains one in WebStorm.