Honestly, the thing that will probably kill LLMs the hardest is someone writing a small language model that fits in JavaScript in a browser and hits comparable benchmarks.

Why bother with all those GPUs and energy usage if your Raspberri Pi could get comparable results?

@soatok i think we're a bunch of innovations away from this goal, it's not implausible. but that's just inference. the part that still sucks time and energy is training. cutting datacenter dependence here and democratizing model construction would truly spell the end of their gatekeeping. here though, i have no ideas.