Learn Claude Code by doing, not reading

https://claude.nagdy.me/

Learn Claude Code Interactively — by Ahmed Nagdy

11 interactive modules take you from first slash command to building plugins — with terminal simulators, config builders, and quizzes.

claude.nagdy.me
Side note: I don’t know what Anthropic changed but now Claude Code consumes the quota incredibly fast. I have the Max5 plan, and it just consumed about 10% of the session quota in 10 minutes on a single prompt. For $100/month, I have higher expectations.
I noticed 1M context window is default and no way not to use it. If your context is at 500-900k tokens every prompt, you’re gonna hit limits fast.

I had to double check that they'd removed the non-1M option, and... WTF? This is what's in `/config` → `model`

1. Default (recommended) Opus 4.6 with 1M context · Most capable for complex work
2. Sonnet Sonnet 4.6 · Best for everyday tasks
3. Sonnet (1M context) Sonnet 4.6 with 1M context · Billed as extra usage · $3/$15 per Mtok
4. Haiku Haiku 4.5 · Fastest for quick answers

So there's an option to use non-1M Sonnet, but not non-1M Opus?

Except wait, I guess that actually makes sense, because it says Sonnet 1M is billed as extra usage... but also WTF, why is Sonnet 1M billed as extra usage? So Opus 1M is included in Max, but if you want the worse model with that much context, you have to pay extra? Why the heck would anyone do that?

The screen does also say "For other/previous model names, specify with --model", so I assume you can use that to get 200K Opus, but I'm very confused why Anthropic wouldn't include that in the list of options.

What a strange UX decision.

`/model opus` sets it to the original non-1M Opus... for now.