Do you hate #broligarchs?
#Billionaires? #AiSlop but still think there is merit in #AI?

Here is my proposal for a stand alone.
OFFGRID COMMUNITY AI SYSTEM.

That's right.Your very own co-op AI

The calculations are very much back of the envelope, first cut, but quite feasible.
A 32billion parameters, frontier level performance compatable open source #llm model. The power requirements is that of 3AC units including cooling. Serves 15-20 concurrent users. 40 households of 4 people each (taking into account actual AI model distributed use metrics and contention ratios)

40 households, subscribing at $30/month over 2 years + power (solar). Train with your own datasets.
Entire set up takes half a rack.

LETS GO!!!

#OpenSource #FOSS #CommunityTech #OpenHardware #EthicalAI #ResponsibleAI #AIForGood #TechForGood #Solarpunk #RegenerativeCulture #Degrowth #AppropriateTechnology #OffGrid #SelfSufficient #Homesteading #Permaculture #RightToRepair #MakerSpace #DIYTech #decentralizedtech

@n_dimension also, I think if we trim down to quality datasets - like Wikipedia and open-source books, I think we can build a smaller model that runs on lower spec hardware.

I run Qwen-3/Jan-code models on my RTX 2060 - no sweat for inference and I can use it for 80-90% of my work. Its like having an interactive encyclopedia offline. I love it.

Specific models for specific use-cases/communities might also be a good idea. Like an agri-trained llm for agriculture

@mahadevank

Google has just released a super tight, great local #LLM I had not had a chance to look at it yet

I really like your agri model idea.

I was thinking a basic medical (nurse level) one for the third world/post-collapse.

@n_dimension @mahadevank I recall there was some research a while back which showed that domain specific fine tuning really did not work well.

There was attempts at training astronomy specific models, and while they outperformed models of a similar size at questions like "describe the lightcurve of binary star mergers" they suffered from much higher hallucination rates, and performed worse at generalising outside of the specific documents they were fine tuned on.

Now admittedly, this was back in the Llama2 days so maybe "modern" architectures would behave differently. But it seems that a broad dataset is necessary for generalising, even within a specific domain

@n_dimension @mahadevank (con't)

For example, there is other research easily available which shows that including programming in the training data MASSIVELY improved performance in mathematics and general problem solving.

@n_dimension @mahadevank (con't again, sorry I started rambling)

That is not to say that general purpose systems are the best and specialised systems won't work.

There was some work apparently by Yann LeCunn (I haven't read it myself yet though) and apparently the optimal architecture game playing AIs was small LLMs combined with domain specific tools

@AuntyRed @n_dimension I'm a total newbie to LLMs and such. Just know a few things about basic neural networks, so not much idea on what makes LLMs tick.

I saw one model though - shellm - which was just a 378M parameter model that was great at writing out shell commands for a given prompt

@mahadevank @n_dimension oh yeah, there's a whole lot you can do with a small model, especially when combined with an external trusted source of data, e.g. a tool to search Wikipedia

@mahadevank @AuntyRed

The models push to prod every week a .. release

However, I don't doubt the research. Multidimensional vector trees that underpin #LLM s have some very peculiar traversal patterns.

To be honest, I did not have enough time to look into the local models. There is a local education RAG hybrid I stumbled upon, that actually looks real solid. The purdy picture is all I have tho.

@n_dimension do you have the name, I'd love to give it a go

@mahadevank

#Gemma4 it comes in 4 sizes with the biggest at 32B parameters and apparently it runs pretty decent on...a mobile device!!!

Apache license.
Sounds like #LLAMA2 #Quen #deepseek killer.

Let me know if you remember its utility, please.
Its literally just out.

@n_dimension nice! I'll get that downloaded - I use llama-cli for the models, works great. llama-server does some funny stuff and fails to allocate a ton of memory, no idea why

@n_dimension gemma-4 didnt work on llama-cli. Might give it a try later but honestly, i'm very happy with the models i have on local

The RAG thing, i'm not sold on - most of them dont send the right slices of context to the LLM so you get weird responses. I'd rather train a new model.

Also, the probabilistic nature of all models and therefore whatever thinking etc. happens means human verification at all levels, for sure.