The surge in #semiconductor stocks, driven by #AI, is shifting towards a more heterogeneous #computelandscape beyond #GPUs. While GPUs excel in training and answer inference, #Cerebras#wholewaferchips offer superior speed for #answerinference but face limitations in #training and #agenticinference due to cost and memory constraints. https://stratechery.com/2026/the-inference-shift/?eicker.news #tech #media #news
The Inference Shift

Agentic inference is going to be different than the inference we use today, and it will change compute infrastructure because speed won’t matter when humans aren’t involved.

Stratechery by Ben Thompson
The release of the first widely adopted reasoning model, #o1, marked a #turningpoint in the evolution of #LLMs. An empirical #study using the #OpenRouter platform analysed over 100 trillion tokens of real-world LLM interactions, revealing substantial adoption of #openweightmodels, the popularity of #creativeroleplay and #codingassistance, and the rise of #agenticinference. https://openrouter.ai/state-of-ai?eicker.news #tech #media #news
State of AI 2025: 100T Token LLM Usage Study | OpenRouter

Read OpenRouter's 2025 State of AI report — an empirical 100 trillion token study of real LLM usage, model trends, and developer insights.

OpenRouter