Anthropic just legally threatened Opencode to make them drop support: https://web.archive.org/web/20260221041617/https://github.com/anomalyco/opencode-anthropic-auth/pull/15#issuecomment-3930558874

Archive link because they deleted the repo after to comply with demands.

In short, Anthropic only wants you using their official walled-garden clients to access the models trained on our open source code.

They are not a lesser evil. They are just as evil as OpenAI.

Stop giving these assholes money. Rent or buy hardware to self-host with privacy and freedom. It is not that hard, I promise.

fix: Align Anthropic OAuth requests with Claude Code by deveworld · Pull Request #15 · anomalyco/opencode-anthropic-auth

Summary Normalize Anthropic OAuth requests to match Claude Code's headers, betas, metadata, tool casing, and model IDs. Remove tool_choice and inject metadata.user_id from ~/.claude.json to sa...

GitHub

@lrvick

> Rent or buy hardware to self-host with privacy and freedom. It is not that hard, I promise.

I would really like to do this instead of using Gemini 3.1 / Antigravity. And I have a 5090 to do it with. But what "self-hostable" solution (because of the models, none are FOSS) is actually competitive with those?

@hopeless a 128G Framework AI Max 395+ gives you ~110GB of VRAM and I am getting ~35t/s w/ qwen3.5 122b for reference once switching to 7.0.0 kernel series which is a big perf boost.

https://frame.work/products/desktop-diy-amd-aimax300/configuration/new

Configure Framework Desktop DIY Edition (AMD Ryzen™ AI Max 300 Series)

Choose from AMD and Intel system options, select your preferred memory and storage, operating system, and more customizations. Available in DIY and pre-built configurations.

Framework
@lrvick @hopeless I’m honestly impressed by that pricing! I thought it would be much higher

@nabeards @hopeless Everyone only wants to buy Nvidia, but the AMD stuff is much cheaper if you are willing to use bleeding edge kernel+drivers as support is very recent.

But if you wait to buy hardware when it is plug and play, prepare to pay double.

@lrvick @nabeards

... sure. But clearly, you don't need a H200 to run the inference for SOTA coding assist. And if it works but is slower on a 5090, I will definitely be motivated to spend money to make it work faster.

The actual inference horsepower seems to be becoming less of a blocker

https://research.google/blog/turboquant-redefining-ai-efficiency-with-extreme-compression/

One big missing part is the tight loop eg Antigravity has around the LLM such that, eg if it produces a bad diff, it retries it from scratch at the cost of a few seconds.

TurboQuant: Redefining AI efficiency with extreme compression