Honestly, the thing that will probably kill LLMs the hardest is someone writing a small language model that fits in JavaScript in a browser and hits comparable benchmarks.

Why bother with all those GPUs and energy usage if your Raspberri Pi could get comparable results?

@soatok You know: It strikes me that a lot of what these LLMs are being used for is essentially summarizing text. Sure, not entirely.

It turns out that we were already looking into how to do this well before ChatGPT, but it never took off.

Those models had some procedural guard rails to ensure it was at least somewhat accurately shortening the text, judging by word frequencies. Sure this wasn't actually summarizing, but at least it strove not to lie!