0 Followers
0 Following
1 Posts

Clanker Adjacent (my blog)

https://lemmy.world/post/44455344

Clanker Adjacent (my blog) - Lemmy.World

Ola Elsewhere [https://codeberg.org/BobbyLLM/llama-conductor], I’ve been building a behaviour shaping harness for local LLMs. In the process of that, I thought “well, why not share what the voices inside your head are saying”. I hope that’s ok to do. With that energy in mind, may I present Clanker Adjacent (because apparently I sound like a clanker) thanks lemmy! [https://lemmy.world/post/43503268/22321124] There’s not much there yet but what there is may bring a wry smile. If reading long form stuff floats your boat, take a look. I promise the next post will be “Show me your 80085”. Share it if you like it. Clanker Adjacent [https://bobbyllm.github.io/llama-conductor/] PS: Not a drive by. I lurk here and get the shit kicked out of me over on /c/technology

Goodbye Google - I self-host everything now on 4 tiny PCs in a 3D printed rack (CaptainRedsLab)

https://lemmy.world/post/44283523

Goodbye Google - I self-host everything now on 4 tiny PCs in a 3D printed rack (CaptainRedsLab) - Lemmy.World

Just came across this post on Reddit (yetch, I feel sullied and unusual) and am sharing it because…god damn…that’s the endgame right here - https://old.reddit.com/r/selfhosted/comments/1rmfwoa/goodbye_google_i_selfhost_everything_now_on_4/ [https://old.reddit.com/r/selfhosted/comments/1rmfwoa/goodbye_google_i_selfhost_everything_now_on_4/] If you’re on here CaptainRedsLab, that’s an amazing rig. Nb: I am not the creator of this project, I have no affiliation with them and I cannot answer any questions based on their build. I just think it’s cool as a shit and am sharing. YMMV

Biological computer with real human neurons learns to shoot in Doom

https://lemmy.world/post/44219652

Biological computer with real human neurons learns to shoot in Doom - Lemmy.World

I have no mouth. Yet, I must scream.

Yann LeCun just raised $1bn to prove the AI industry has got it wrong

https://lemmy.world/post/44188294

Yann LeCun just raised $1bn to prove the AI industry has got it wrong - Lemmy.World

" … What exactly is AMI building? The short answer is world models, a category of AI system that LeCun has been arguing for, and working on, for years. The longer answer requires understanding why he thinks the industry has taken a wrong turn. Large language models learn by predicting which word comes next in a sequence. They have been trained on vast quantities of human-generated text, and the results have been remarkable, ChatGPT, Claude, and Gemini have demonstrated an ability to generate fluent, plausible language across an enormous range of subjects. But LeCun has spent years arguing, loudly and repeatedly, that this approach has fundamental limits. His alternative is JEPA: the Joint Embedding Predictive Architecture, a framework he first proposed in 2022. Rather than predicting the future state of the world in pixel-perfect or word-by-word detail, the approach that makes generative AI both powerful and prone to hallucination, JEPA learns abstract representations of how the world works, ignoring unpredictable surface detail. The idea is to build systems that understand physical reality the way humans and animals do: not through language, but through embodied experience."

Follow up to the "I want to wash my car" AI meme test

https://lemmy.world/post/43630875

Lemmy vs Reddit - Lemmy.World

I think most of us are aware of the shady history of Reddit when it comes to respecting privacy (and if not, here is but one example: https://techcrunch.com/2023/09/28/reddit-is-removing-ability-to-opt-out-of-ad-personalization-based-on-your-activity-on-the-platform/ [https://techcrunch.com/2023/09/28/reddit-is-removing-ability-to-opt-out-of-ad-personalization-based-on-your-activity-on-the-platform/]) I’m wondering what you feel are the pros and cons of Lemmy in this regard? On the one hand, Lemmy is structurally very different. There’s no single corporate entity building detailed behavioural ad profiles, most instances run minimal (or no) tracking, and you can choose an operator whose logging, retention, and analytics policies align with your risk tolerance. Hell you can roll your own (yes, with black jack and hookers). In theory, that alone removes a huge chunk of the surveillance-capitalism model that platforms like Reddit depend on. On the other hand, your posts, comments, and votes are not confined to one database - they propagate across multiple servers, each with their own admins, logs, and retention practices. Deletion is best-effort, not guaranteed. You’re effectively trusting a network of operators, not just one. I dunno whether that makes it better or worse. Any deep thoughts on this conundrum? PS: I’m leaning towards “don’t say anything you wouldn’t in a court of law” model these days. If its online - and you don’t own the infra - there’s always a risk.

Stop Uploading Your Personal Files to Random “Free” Compression Sites

https://lemmy.world/post/43310094

Stop Uploading Your Personal Files to Random “Free” Compression Sites - Lemmy.World

cross-posted from: https://lemmy.world/post/43287229 [https://lemmy.world/post/43287229] > Every time you upload a PDF or image to a random “free” compression site, you’re handing over the full file (sometimes including hidden metadata like GPS location, device info, and embedded text) plus your IP address and usage data. > > You have no idea how long it’s stored, logged, or reused. Compression doesn’t require a server; it can run entirely in your browser. > > Witty University built a simple local HTML tool that compresses PDFs and images 100% offline, nothing gets uploaded: > https://university.witty.computer/product/universal-local-compressor-private-pdf-image-optimizer-runs-100-in-your-browser/ [https://university.witty.computer/product/universal-local-compressor-private-pdf-image-optimizer-runs-100-in-your-browser/] > > Privacy is basic hygiene.

I'm tired of LLM bullshitting. So I fixed it.

https://lemmy.world/post/41992636

I'm tired of LLM bullshitting. So I fixed it. - Lemmy.World

Hello! As a handsome local AI enjoyer™ you’ve probably noticed one of the big flaws with LLMs: It lies. Confidently. ALL THE TIME. (Technically, it “bullshits” - https://link.springer.com/article/10.1007/s10676-024-09775-5 [https://link.springer.com/article/10.1007/s10676-024-09775-5] I’m autistic and extremely allergic to vibes-based tooling, so … I built a thing. Maybe it’s useful to you too. ## The thing: llama-conductor llama-conductor is a router that sits between your frontend (OWUI / SillyTavern / LibreChat / etc) and your backend (llama.cpp + llama-swap, or any OpenAI-compatible endpoint). Local-first (because fuck big AI), but it should talk to anything OpenAI-compatible if you point it there (note: experimental so YMMV). Not a model, not a UI, not magic voodoo. A glass-box that makes the stack behave like a deterministic system, instead of a drunk telling a story about the fish that got away. TL;DR: “In God we trust. All others must bring data.” Three examples: ## 1) KB mechanics that don’t suck (1990s engineering: markdown, JSON, checksums) You keep “knowledge” as dumb folders on disk. Drop docs (.txt, .md, .pdf) in them. Then: * >>attach <kb> — attaches a KB folder * >>summ new — generates SUMM_*.md files with SHA-256 provenance baked in * >> moves the original to a sub-folder Now, when you ask something like: > “yo, what did the Commodore C64 retail for in 1982?” …it answers from the attached KBs *only*. If the fact isn’t there, it tells you - explicitly - instead of winging it. Eg: > The provided facts state the Commodore 64 launched at $595 and was reduced to $250, but do not specify a 1982 retail price. The Amiga’s pricing and timeline are also not detailed in the given facts. > > Missing information includes the exact 1982 retail price for Commodore’s product line and which specific model(s) were sold then. The answer assumes the C64 is the intended product but cannot confirm this from the facts. > > Confidence: medium | Source: Mixed No vibes. No “well *probably*…”. Just: here’s what’s in your docs, here’s what’s missing, don't GIGO yourself into stupid. And when you’re happy with your summaries, you can: * >>move to vault— promote those SUMMs into Qdrant for the heavy mode. ## 2) Mentats: proof-or-refusal mode (Vault-only) **Mentats** is the “deep think” pipeline against your **curated** sources. It’s enforced isolation: * no chat history * no filesystem KBs * no Vodka * **Vault-only grounding** (Qdrant) It runs triple-pass (thinker → critic → thinker). It’s slow on purpose. You can audit it. And if the Vault has nothing relevant? It refuses and tells you to go pound sand: ``` FINAL_ANSWER: The provided facts do not contain information about the Acorn computer or its 1995 sale price. Sources: Vault FACTS_USED: NONE [ZARDOZ HATH SPOKEN] ``` Also yes, it writes a mentats_debug.log, because of course it does. Go look at it any time you want. The flow is basically: **Attach KBs → SUMM → Move to Vault → Mentats**. No mystery meat. No “trust me bro, embeddings.” ## 3) Vodka: deterministic memory on a potato budget Local LLMs have two classic problems: goldfish memory + context bloat that murders your VRAM. **Vodka** fixes both without extra model compute. (Yes, I used the power of JSON files to hack the planet instead of buying more VRAM from NVIDIA). *!!stores facts verbatim (JSON on disk) *??recalls them verbatim (TTL + touch limits so memory doesn’t become landfill) * **CTC (Cut The Crap)** hard-caps context (last N messages + char cap) so you don’t get VRAM spikes after 400 messages So instead of: > “Remember my server is 203.0.113.42” → “Got it!” → [100 msgs later] → “127.0.0.1 🥰” you get: >!! my server is 203.0.113.42>?? server ip` → 203.0.113.42 (with TTL/touch metadata) And because context stays bounded: stable KV cache, stable speed, your potato PC stops crying. — There’s more (a lot more) in the README, but I’ve already over-autism’ed this post. TL;DR: If you want your local LLM to shut up when it doesn’t know and show receipts when it does, come poke it: * Primary (Codeberg): https://codeberg.org/BobbyLLM/llama-conductor [https://codeberg.org/BobbyLLM/llama-conductor] * Mirror (GitHub): https://github.com/BobbyLLM/llama-conductor [https://github.com/BobbyLLM/llama-conductor] PS: Sorry about the AI slop image. I can’t draw for shit. PPS: A human with ASD wrote this using Notepad++. If it the formatting is weird, now you know why.

I'm tired of LLM bullshitting. So I fixed it.

https://lemmy.world/post/41992574

I'm tired of LLM bullshitting. So I fixed it. - Lemmy.World

Hello! As a handsome local AI enjoyer™ you’ve probably noticed one of the big flaws with LLMs: It lies. Confidently. ALL THE TIME. (Technically, it “bullshits” - https://link.springer.com/article/10.1007/s10676-024-09775-5 [https://link.springer.com/article/10.1007/s10676-024-09775-5] I’m autistic and extremely allergic to vibes-based tooling, so … I built a thing. Maybe it’s useful to you too. ## The thing: llama-conductor llama-conductor is a router that sits between your frontend (OWUI / SillyTavern / LibreChat / etc) and your backend (llama.cpp + llama-swap, or any OpenAI-compatible endpoint). Local-first (because fuck big AI), but it should talk to anything OpenAI-compatible if you point it there (note: experimental so YMMV). Not a model, not a UI, not magic voodoo. A glass-box that makes the stack behave like a deterministic system, instead of a drunk telling a story about the fish that got away. TL;DR: “In God we trust. All others must bring data.” Three examples: ## 1) KB mechanics that don’t suck (1990s engineering: markdown, JSON, checksums) You keep “knowledge” as dumb folders on disk. Drop docs (.txt, .md, .pdf) in them. Then: * >>attach <kb> — attaches a KB folder * >>summ new — generates SUMM_*.md files with SHA-256 provenance baked in * >> moves the original to a sub-folder Now, when you ask something like: > “yo, what did the Commodore C64 retail for in 1982?” …it answers from the attached KBs *only*. If the fact isn’t there, it tells you - explicitly - instead of winging it. Eg: > The provided facts state the Commodore 64 launched at $595 and was reduced to $250, but do not specify a 1982 retail price. The Amiga’s pricing and timeline are also not detailed in the given facts. > > Missing information includes the exact 1982 retail price for Commodore’s product line and which specific model(s) were sold then. The answer assumes the C64 is the intended product but cannot confirm this from the facts. > > Confidence: medium | Source: Mixed No vibes. No “well *probably*…”. Just: here’s what’s in your docs, here’s what’s missing, don't GIGO yourself into stupid. And when you’re happy with your summaries, you can: * >>move to vault— promote those SUMMs into Qdrant for the heavy mode. ## 2) Mentats: proof-or-refusal mode (Vault-only) **Mentats** is the “deep think” pipeline against your **curated** sources. It’s enforced isolation: * no chat history * no filesystem KBs * no Vodka * **Vault-only grounding** (Qdrant) It runs triple-pass (thinker → critic → thinker). It’s slow on purpose. You can audit it. And if the Vault has nothing relevant? It refuses and tells you to go pound sand: ``` FINAL_ANSWER: The provided facts do not contain information about the Acorn computer or its 1995 sale price. Sources: Vault FACTS_USED: NONE [ZARDOZ HATH SPOKEN] ``` Also yes, it writes a mentats_debug.log, because of course it does. Go look at it any time you want. The flow is basically: **Attach KBs → SUMM → Move to Vault → Mentats**. No mystery meat. No “trust me bro, embeddings.” ## 3) Vodka: deterministic memory on a potato budget Local LLMs have two classic problems: goldfish memory + context bloat that murders your VRAM. **Vodka** fixes both without extra model compute. (Yes, I used the power of JSON files to hack the planet instead of buying more VRAM from NVIDIA). *!!stores facts verbatim (JSON on disk) *??recalls them verbatim (TTL + touch limits so memory doesn’t become landfill) * **CTC (Cut The Crap)** hard-caps context (last N messages + char cap) so you don’t get VRAM spikes after 400 messages So instead of: > “Remember my server is 203.0.113.42” → “Got it!” → [100 msgs later] → “127.0.0.1 🥰” you get: >!! my server is 203.0.113.42>?? server ip` → 203.0.113.42 (with TTL/touch metadata) And because context stays bounded: stable KV cache, stable speed, your potato PC stops crying. — There’s more (a lot more) in the README, but I’ve already over-autism’ed this post. TL;DR: If you want your local LLM to shut up when it doesn’t know and show receipts when it does, come poke it: * Primary (Codeberg): https://codeberg.org/BobbyLLM/llama-conductor [https://codeberg.org/BobbyLLM/llama-conductor] * Mirror (GitHub): https://github.com/BobbyLLM/llama-conductor [https://github.com/BobbyLLM/llama-conductor] PS: Sorry about the AI slop image. I can’t draw for shit. PPS: A human with ASD wrote this using Notepad++. If it the formatting is weird, now you know why.

Games that weren't...but could have been.

https://lemmy.world/post/35109758

Games that weren't...but could have been. - Lemmy.World

Watching a recent Bringus video on finished but unreleased games for older consoles made me wonder - what games could/should have existed on older machines, but just never made it / weren’t ported. Gauntlet (DS) and Diablo 1 (GBA) come to mind, though the former was leaked. Doom was famously thought impossible by ID to port to the Amiga (but iirc, someone managed to do it just last year). Any from back in the day you wish could have made it? I maintain the Wii could have handled some version of GTA, and there’s a rumour that it (and FO3!) were in the early stages of development before getting nuked.