Kann krankheitsbedingt nicht zum #tpus

a blue cartoon character is cr...
\o/ #tpus Tag \o/
Verdammt. Ich glaube, ich kriege Schnupfen. Wehe. Ich will Donnerstag zum #tpus !! (Und überhaupt)
Mein Wochenplan:
Heute: Lesung
Morgen: Lesung
Mittwoch: Konzert
Donnerstag: #tpus
Freitag: Nix
Sonnabend: Lesung
#Nvidia secured a non-exclusive licensing agreement with #Groq, an #AIchip startup, for $20 billion. The deal aims to bring Groq’s CEO, #JonathanRoss, on board, along with their #inferencetechnology and #intellectualproperty. This move is seen as a strategic move to counter #Google’s success with #TPUs and maintain Nvidia’s dominance in the #AIchipmarket. https://spyglass.org/nvidia-groq-deal/?eicker.news #tech #media #news
NVIDIA’s Christmas Eve 'Hackquisition' Miracle

NVIDIA gets a creator of the TPU, access to inference tech & IP, and stops anyone else from getting such things. All for a mere $20B...

Spyglass
Touching the Elephant - TPUs

Understanding the Tensor Processing Unit

Consider the Bulldog

"If you go back a year or two, you might make the case that Nvidia had three moats relative to TPUs: superior performance, significantly more flexibility due to GPUs being more general purpose than TPUs, and CUDA and the associated developer ecosystem surrounding it. OpenAI, meanwhile, had the best model, extensive usage of their API, and the massive number of consumers using ChatGPT.

The question, then, is what happens if the first differentiator for each company goes away? That, in a nutshell, is the question that has been raised over the last two weeks: does Nvidia preserve its advantages if TPUs are as good as GPUs, and is OpenAI viable in the long run if they don’t have the unquestioned best model?

Nvidia’s flexibility advantage is a real thing; it’s not an accident that the fungibility of GPUs across workloads was focused on as a justification for increased capital expenditures by both Microsoft and Meta. TPUs are more specialized at the hardware level, and more difficult to program for at the software level; to that end, to the extent that customers care about flexibility, then Nvidia remains the obvious choice.

CUDA, meanwhile, has long been a critical source of Nvidia lock-in, both because of the low level access it gives developers, and also because there is a developer network effect: you’re just more likely to be able to hire low level engineers if your stack is on Nvidia. The challenge for Nvidia, however, is that the “big company” effect could play out with CUDA in the opposite way to the flexibility argument. While big companies like the hyperscalers have the diversity of workloads to benefit from the flexibility of GPUs, they also have the wherewithal to build an alternative software stack. That they did not do so for a long time is a function of it simply not being worth the time and troube..."

https://stratechery.com/2025/google-nvidia-and-openai/

#AI #GenerativeAI #Nvidia #Google #ChatGPT #OpenAI #LLMs #Chatbots #CUDA #GPUs #TPUs

Google, Nvidia, and OpenAI

OpenAI and Nvidia are both under threat from Google; I like OpenAI’s chances best, but they need an advertising model to beat Google as an Aggregator.

Stratechery by Ben Thompson

"In the blistering race for AI supremacy, Nvidia has long reigned as the undisputed king. Its GPUs powered the explosive growth of machine learning, turning abstract neural networks into reality and fueling an empire valued at trillions. But as the AI landscape evolves, cracks are appearing in Nvidia's armor. The shift from model training (Nvidia's stronghold) to inference, the real-time application of those models, is reshaping the market. And at the forefront of this revolution stands Google's Tensor Processing Units (TPUs), delivering unmatched efficiency and cost savings that could spell the end of Nvidia's monopoly.

By 2030, inference will consume 75% of AI compute, creating a $255 billion market growing at 19.2% annually. Yet most companies still optimize for training costs. This isn't just hype; it's economics. Training is a one-time sprint, but inference is an endless marathon. As companies like OpenAI grapple with skyrocketing inference bills (projected at $2.3 billion for 2024 alone, dwarfing the $150 million cost to train GPT-4), Google's TPUs emerge as the cost-effective powerhouse. In this in-depth analysis, we'll explore how TPUs are winning the inference war, backed by real-world migrations from industry leaders, and why this pivot signals Nvidia's impending decline."

https://www.ainewshub.org/post/ai-inference-costs-tpu-vs-gpu-2025

#AI #AIInference #GenerativeAI #Nvidia #Google #GPUs #TPUs #LLMs

AI Inference Costs 2025: Why Google TPUs Beat Nvidia GPUs by 4x

Nvidia's AI empire is crumbling. Google's TPUs now deliver 4x better performance-per-dollar for inference, the workload consuming 75% of AI compute by 2030. Midjourney slashed costs 65% by switching. Meta's negotiating multibillion-dollar TPU deals. Even Wall Street legends like Peter Thiel and Michael Burry are dumping $6B+ in Nvidia stock. The inference era has arrived, and specialized ASICs are winning.

AI News Hub
#Google is selling its #TPU #chips to external customers, challenging #Nvidia’s dominance in the #AIhardware market. #Anthropic, a major AI company, is a significant customer, purchasing over 1 million #TPUs. This move positions Google as a direct competitor to Nvidia, offering a differentiated #cloudprovider option with its #inhousesilicon design capabilities. https://newsletter.semianalysis.com/p/tpuv7-google-takes-a-swing-at-the?eicker.news #tech #media #news
Google TPUv7: The 900lb Gorilla In the Room

Anthropic’s 1GW+ TPUs, New customers Meta/SSI/xAI/OAI, Full Stack Review of v7 Ironwood, CUDA Moat at risk, Next Generation TPUv8AX and TPUv8X versus Vera Rubin

SemiAnalysis

TPUs vs. GPUs and why Google is positioned to win AI race in the long term

https://www.uncoveralpha.com/p/the-chip-made-for-the-ai-inference

#HackerNews #TPUs #GPUs #Google #AI #race #technology

The chip made for the AI inference era – the Google TPU

I am publishing a comprehensive deep dive, not just a technical overview, but also strategic and financial coverage of the Google TPU.

UncoverAlpha