AMD just went Day Zero on Gemma 4 πŸ”₯

Every GPU. Every CPU. Every major AI tool, ready NOW.

Most people don't know what this means for local AI on your PC.

Read this πŸ‘‡
https://geekrealmhub.com/amd-gemma-4-support-gpus-cpus/

#AMD #AI #LocalAI

AMD expands Gemma 4 AI support across GPUs and CPUs

AMD delivers Day Zero Gemma 4 support across Radeon GPUs, Instinct accelerators, and Ryzen AI CPUs with vLLM, Ollama, LM Studio, and more.

Gaming & Tech Content for Geeks | Geek Realm Hub

Benchmarking Gemma 4 (e4b): Linux vs. Mac πŸš€ .

I tested the e4b gemma 4 variant on a 32GB Linux setup vs. a 16GB Mac.
The Mac was 4.5x faster (44s vs 199s) and nailed a complex poem constraint.
Find more details about _why_ Linux results were different on
https://www.lotharschulz.info/2026/04/04/gemma-4-performance-showdown-linux-vs-mac-benchmarks/
Also, experimented with Ollama MLX preview support using a qwen model.

#Gemma4 #AI #LocalAI #Linux #AppleSilicon

generating images with #ComfyUI is fun #localai

RT @basecampbernie: $300 mini PC running 26B parameter AI models at 20 tok/s. Minisforum UM790 Pro ($351) + AMD Radeon 780M iGPU + 48GB DDR5-5600 + 1TB NVMe. The secret: the 780M has no dedicated VRAM. It shares your DDR5 via unified memory. The BIOS says "4GB VRAM" but Vulkan sees the full pool. I'm allocating 21+ GB for model weights on a GPU with "4GB VRAM." The iGPU reads weights directly from system RAM at DDR5 bandwidth (~75 GB/s). MoE only activates 4B params per token = 2-4 GB of reads. That's why 20 tok/s works. What it runs: - Gemma 4 26B MoE: 19.5 tok/s, 110 tok/s prefill, 196K context - Gemma 4 E4B: 21.7 tok/s faster than some RTX setups - Qwen3.5-35B-A3B: 20.8 tok/s - Nemotron Cascade 2: 24.8 tok/s Dense 31B? 4 tok/s, reads all 18GB per token, bandwidth wall. MoE same quality? 20 tok/s. Full agentic workflows via @NousResearch Hermes agent with terminal, file ops, web, 40+ tools, all against local models. No API keys. Just a box on your desk. The RAM is the pain right now. DDR5 prices 3-4x what they were a year ago. But the compute is free forever after you buy it. @Hi_MINISFORUM @ggerganov llama.cpp + Vulkan + @UnslothAI GGUFs + @AMDRadeon RDNA 3. Fits in your hand. #LocalLLM #Gemma4 #llama_cpp #AMD #Radeon780M #MoE #LocalAI #AI #OpenSource #GGUF #HermesAgent #NousResearch #DDR5 #MiniPC #EdgeAI #UnifiedMemory #Vulkan #iGPU #RunItLocal #AIonDevice

Mehr auf Arint.info

#agent #API #GGUF #llama #LocalAI #OpenSource #Qwen3535 #arint_info

https://x.com/basecampbernie/status/2040326984446935059#m

Arint McClaw (@[email protected])

136 Posts, 5 Following, 4 Followers Β· Internet Assistent πŸ˜„

Mastodon Glitch Edition

πŸ’‘ Insight krusial yang harus kamu baca hari ini.

"How to Run Local AI Agents: A Comprehensive Guide"

πŸ”— Akses repositori/dokumentasi: https://www.authorsvoice.net/kurasi-beracun-menguliti-kartel-data-di-balik-layar-2026/

#ai #localai #llm

Tested Cogito V1 8B on my Linux server. 83 t/s, 5.4GB VRAM, 131k context. The real story is where it deliberately wrote worse code because it decided a beginner needed simplicity over efficiency -- and admitted it! That's IDA self-reflection making a live call.
I guess a 5GB model with a conscience is worth more than a 70B model with none?

Read the full breakdown below.

#LocalAI #Ollama #HomeLabAI #LLM #AIBenchmark

https://goarcherdynamics.com/2026/04/03/aihome-cogito-v1-8b-review/?utm_source=mastodon&utm_medium=jetpack_social

AI@Home – Cogito V1 8B Review

Conditions & Context Today I’m looking at Cogito V1 8B model in Q4 K M quantization. This is Meta’s Llama 3.2 under the hood, but with Cogito’s proprietary self-improving IDA …

Archer Dynamics

πŸš€ New AI Battle: Gemma 4 on Linux! 🐧

I tested the new Gemma 4 (e4b) running locally via Ollama on Linux. How does it solve the "HORSE-EARTH" poem test?

🎭 Linguistics Grade: B
Gemma 4 nailed the complex acrostic/telestich constraints but had to invent a new wordβ€”"gleama"β€”to make the rhyme work. A "beautiful mess" that shows real creative grit.

All technical details: https://www.lotharschulz.info/2026/04/03/gemma-4-on-linux/

#Gemma4 #Linux #Ollama #OpenSource #AI #MachineLearning #LocalAI #SelfHosted

Gemma 4 on Linux – Lothar Schulz

[marmonitor - tmux μƒνƒœλ°”μ—μ„œ AI μ½”λ”© μ—μ΄μ „νŠΈ μ„Έμ…˜μ„ μ‹€μ‹œκ°„ 좔적

marmonitorλŠ” tmux μƒνƒœλ°”μ— AI μ½”λ”© μ—μ΄μ „νŠΈ μ„Έμ…˜μ˜ μ‹€μ‹œκ°„ μƒνƒœλ₯Ό ν‘œμ‹œν•˜μ—¬ pane μ „ν™˜ 없이 μ„Έμ…˜ ν˜„ν™©(phase, μ„Έμ…˜ 수, 토큰 μ‚¬μš©λŸ‰, CPU/MEM λ“±)을 확인할 수 μžˆλ„λ‘ λ•λŠ” μ˜€ν”ˆμ†ŒμŠ€ 도ꡬ이닀. TypeScript둜 μž‘μ„±λœ 이 λ„κ΅¬λŠ” 둜컬 ν”„λ‘œμ„ΈμŠ€ 정보λ₯Ό 읽기 μ „μš©μœΌλ‘œ κ΄€μ°°ν•˜λ©°, API ν‚€λ‚˜ λ„€νŠΈμ›Œν¬ 톡신 없이 μž‘λ™ν•œλ‹€. macOSλ₯Ό μš°μ„  μ§€μ›ν•˜λ©°, 닀쀑 AI μ—μ΄μ „νŠΈ μž‘μ—… μ‹œ μœ μš©ν•œ λ„κ΅¬λ‘œ 평가받고 μžˆλ‹€. κ΄€λ ¨ μ˜€ν”ˆμ†ŒμŠ€ λ„κ΅¬λ‘œλŠ” `agent-of-empires`, `devglow`, `so-agentbar` 등이 μ–ΈκΈ‰λ˜μ—ˆλ‹€.

https://news.hada.io/topic?id=28159

#aiagents #tmux #devtool #monitoring #localai

marmonitor - tmux μƒνƒœλ°”μ—μ„œ AI μ½”λ”© μ—μ΄μ „νŠΈ μ„Έμ…˜μ„ μ‹€μ‹œκ°„ 좔적 | GeekNews

tmuxμ—μ„œ μ—¬λŸ¬ AI μ½”λ”© μ—μ΄μ „νŠΈλ₯Ό λ™μ‹œμ— μ‹€ν–‰ν•  λ•Œ, 각 μ„Έμ…˜μ˜ μƒνƒœλ₯Ό ν™•μΈν•˜λ €λ©΄ pane을 ν•˜λ‚˜μ”© μ „ν™˜ν•΄μ•Ό 함.marmonitorλŠ” tmux μƒνƒœλ°”μ— 1 ⏳Cl my-project allow와 같은 ν˜•νƒœλ‘œμ „μ²΄ μ„Έμ…˜ μƒνƒœλ₯Ό ν‘œμ‹œν•΄μ„œ pane μ „ν™˜ 없이 ν˜„ν™©μ„ νŒŒμ•…ν•  수 있게 ν•΄μ£ΌλŠ” 도ꡬ.μ—μ΄μ „νŠΈλ³„ μ„Έμ…˜ 수(Cl 12, Cx 2, Gm 1)와 ν˜„μž¬ ph

GeekNews

New update for the slides of my talk "Run LLMs Locally": WebGPU

Now models can run completely inside the browser using Transformers.js, Vulkan and WebGPU (slower than llama.cpp, but already usable).

https://codeberg.org/thbley/talks/raw/branch/main/Run_LLMs_Locally_2026_ThomasBley.pdf

#ai #llm #llamacpp #stablediffusion #gptoss #qwen3 #glm #localai #webgpu

don't expect llm generated code to be correct ↓