here we go!
AI counter within 2 sentences? 2
here we go!
AI counter within 2 sentences? 2
LMAO, freudian slip:
AI Error instead of Era
Mi300X is "launching"
TF32 FP8, sparsity support on top of fp/bf16 and ofc int8
Dozen chiplets (including HBM stacks?)
4 io dies with vcache built in (!!!!)
8 GPU tiles
Some AI inference benchmarks, kind of cool but not super shocking.
comparison with H100 (not surprising) but back of the envelop puts Mi300 as within a few dozen percentage points of H200 when both in an SPR host machine with 8x GPGPU
building the foundation of AI for "x" years? sure.
We've been waiting/working on them since the fifth gen language discussions of the 80 and 90's
DAT DENSITY
OCP compliant is pretty pog NGL
"Open, proven and ready" is one way to put ROCm 😬
Love you guys and galls, but I hard disagree with the claim
ROCm 6: adds support for FP8 (which ones?)
Graph compiler is cool but not that new.
still good to see as long as it gets *delivered*
Software partners is good, running natively* with pytorch and requiring no code is also good.
native support for Mi300 in trident from OpenAI
Meh, how many times has "software" 2.0 been brought up?
first with high level langs, then run time langs etc.
will "Ai" change some parts of software developement, probably?
"reshape" the enterprise, maybe? but every "new" thing has changed some parts of the enterprise, question is always how much
Back to the real presentation
ROCm 6 is "SoonTM" (next year? next quarter? etc.)
Ehhhhhhh dell getting Mi300! (knew it was coming, still good to see)
8xMi300X
New plan for smart doorbells:
use different voice clips of VPs, principal engineers, CEOs etc. for saying "good morning"
opening up of infinity fabric is super cool.
I assume this is GPGPU infinity fabric, not chiplet to chiplet IF?
( @philparkbot ?)
MI300A in volume
4 IO dies (256 MB of L3 cache, also proably how they're buffering CPU to GPU cache coherence?)
3 Zen4 CCDs
6 GPU CCD
A little sad that Mi300A is "only" getting HBM3 and not 3e
Sort of expect it, but Mi300X is getting the "fun stuff"
IIRC el cap is ~10% public science?
Either way, el cap will be an interesting machine
Ok so we're now calling it "AI tops"
not a terrible name for int8?
39 ai tops I think is number one unless you go to dGPUs?
Even above Apples offerings IIRC
Clarification here, that's across the SoC, not just the NPU.
so would be number 2 to Apple
Still time for a "one more thing"
please do it lisa 🥺 gib Mi300C
Next time!
That's all folks!
@hattom yeah, Mi300A will be comparable to Grace Hopper *100*
Mi300X is against H200 (none Grace version)