@kuketzblog Die Wende kommt mit #slm small language modell. Kleine lokale hochspezialisierte Modelle die die Entwickler nur beim sicheren coden unterstützen sowie die es benötigen. Ohne cloud ohne proprietäre Kacke ohne big tech.Es ist nur ein Frage der Zeit. Dann platzt die KI Blase. Zu begin war das Mainframe das Maß aller Dinge bis einer mir der Idee ankam das jeder einen Computer haben könnte. #justmy2cents
You want to download a #browser to surf the web. Everyone recommends #Google #Chrome.
You download it and get shocked. What, 4,7 GB for a browser?
Well, it comes with something you really desperately want. Something you cannot live without.
A #SLM (small language model).
You just wanted to surf the internet and now you get #AI, if you want or not. The same with #Microsoft #Edge.
#iOS, #macOS, #Android all have local models in their OS or you get it with the OS, if you want it or not.
The last places left? #GrapheneOS for your phone, #Linux for your computer.
#Enshittification at the speed of AI
Small language models: Rethinking enterprise AI architecture

As LLMs hit the limits of scale and cost, specialized SLMs are emerging as the faster, cheaper, and more private workhorse for the autonomous enterprise.

InfoWorld

Small language models do not just shrink the cloud. They end it. The sovereignty thesis becomes practical the day the model fits on the device.

https://mickai.co.uk/articles/small-language-models-finally-unlock-sovereign-ai

#smalllanguagemodels #slm #ondeviceai

Small language models do not just shrink the cloud. They end it. The sovereignty thesis becomes practical the day the model fits on the device.

The early-2026 SLM wave (Phi-4, Llama 3.3 8B, Gemma 3 4B, Mistral Small 3, Qwen 2.5 3B) is being sold as cheaper inference. The real story is sovereignty. With the model on the device, the vendor leaves the trust chain, and a per-action signed audit ledger becomes feasible for the first time. Mickai's filed-patent stack was waiting for this moment.

Posting some light LLM research I've worked on in my spare time. It doesn't count if it's not on your blog!

Also nanochat was pretty straight forward to build on top of. Good for model hacking.

https://dvshkn.com/blog/2026-05-01-implementing-transformer-model-growth-via-layer-stacking-in-nanochat/

#llm #slm #machinelearning #deeplearning

Implementing Transformer Model Growth via Layer Stacking in Nanochat - Davetown

#ollama #LLM #SLM #Hallucinations #Smollm #LocalLLM incredible insights from a 135M parameter model

TestingCatalog News (@testingcatalog)

Plurai가 에이전트를 위한 실시간 맞춤형 eval과 가드레일을 빠르게 구축하는 ‘vibe-training’을 소개했다. 의도에서 프로덕션용 API 엔드포인트까지 몇 분 만에 만들 수 있고, SLM은 100ms 미만 지연과 8배 이상 저렴한 비용을 제공한다고 강조했다.

https://x.com/testingcatalog/status/2049145106620367065

#aiagent #evaluation #guardrails #slm #latency

TestingCatalog News 🗞 (@testingcatalog) on X

Plurai introduced vibe-training 👀 A new way to build real-time, tailored evals and guardrails for your agent, with high accuracy at a fraction of the LLM cost. > Goes from intent to a production-ready API endpoint in minutes > SLMs run at sub-100ms latency, over 8x cheaper

X (formerly Twitter)

Fiz uma análise para o @outraspalavras.net de documento da Comissão de Análise Econômica e de Segurança EUA-China sobre como o país asiático escolheu disputar a corrida industrial por IA e o que o Sul Global pode aprender com isso.

https://outraspalavras.net/tecnologiaemdisputa/ia-como-a-china-esta-vencendo/

#AI #China #US #data #SLM

Da will ich eine alte Nvidia 3060 reaktivieren, um noch etwas mit #SLM / #LLM zu testen - und habe ausversehen die uralte 660 eingebaut.

Wusste gar nicht mehr, dass ich die noch hatte.

Warum müssen Grafikkarten erstmal so ähnlich aussehen?

Ich bin inzwischen in der "Ich lasse ein gutes #LLM das Arbeitsergebnis von mehreren #SLM s evaluieren."-Phase angekommen...