Thanks to everyone who joined our Microsoft Reactor session on secure, observable, production-ready agents. Repo + slides for the demo are here:
https://aka.ms/microsoftnvidiademo
Thanks to everyone who joined our Microsoft Reactor session on secure, observable, production-ready agents. Repo + slides for the demo are here:
https://aka.ms/microsoftnvidiademo
Production-ready agents need more than a prompt loop. Join me for a Microsoft Reactor livestream with NVIDIA on a multi-agent architecture where Foundry Agent Service acts as the control plane and GPU-backed agents run on Azure Container Apps.
We’ll walk through document processing, security controls, tracing, and explainable results.
🗓 Mar 11, 2026 • 9 AM PT / 6 PM PT
👉 https://developer.microsoft.com/en-us/reactor/events/26660/
Endre Stølsvik (@stolsvik)
DeepSeek이 자사 AI 네트워크, 학습 데이터, 인프라를 완전히 오픈소스로 공개했다는 주제의 트윗입니다. Anthropic이 이를 참고하거나 도구를 활용했다는 주장도 있습니다. AI 모델 개발 투명성과 오픈소스 전략의 파급력을 보여줍니다.

This is so true: @DeepSeek_ai give out *all* their secrets, network build, training, even infra and systems *as Open Source*. @AnthropicAI has *obviously* read it all, taken inspiration, maybe even straight up *used* their tooling. Then they whine when DS loans a bit back. 🤬
FutureLivingLab (@FutureLab2025)
LLM에서 가장 어려운 부분은 모델 자체가 아니라 인프라라는 관점의 글. Andrej Karpathy가 언급한 'vibe coding'처럼 빠르게 개발하는 방식은 초기 속도에는 도움이 되지만 AI 인프라에서는 확장성이 떨어져 '빨리 배포 → 더 빨리 리팩터'하는 문제로 이어진다는 교훈과 대규모 시스템에서 먼저 깨지는 부분들에 대한 논의를 시작한다.

The hardest part of LLMs isn’t the model — it’s the infra. As @karpathy has discussed, “vibe coding” can be a great way to move fast. Our lesson from AI infra: vibes alone don’t scale — they turn into “ship fast → refactor faster”. What breaks first in large systems: ①
Neocloud Economics: CoreWeave vs Nebius – Vertical AI Infra Crushes Hyperscalers (60-70% Margins) ⚡
Neoclouds own stack (chips→racks), dodge AWS debt/leasing. Nebius edges CoreWeave on costs; $10B+ ARR potential. AI training explodes demand
Why vertical? 2-3x cheaper GPUs vs cloud giants.
VCs: Next hyperscalers? Founders: Build atop. GPU wars incoming. 📈
⚡ Cut GPU costs by 68% without slowing inference.
No hype. Just real infra lessons from shipping AI in production.
👉 Read the full story:
https://medium.com/@rogt.x1997/why-my-ai-startup-cut-gpu-costs-by-68-without-slowing-down-with-runpod-58f71733e86f
#GenAI #AIInfra #MLOps
https://medium.com/@rogt.x1997/why-my-ai-startup-cut-gpu-costs-by-68-without-slowing-down-with-runpod-58f71733e86f
ARBITER: what it is / what it isn’t
IS
semantic scoring
geometric fit
negative answers
offline 26MB
ISN’T
LLM
vector DB
embeddings
retrieval
getarbiter.dev
#AI #NLP #RAG #AIInfra #SemanticSearch