https://www.kimi.com/blog/kimi-vendor-verifier
This column is a good survey of where we're at right now and is written by someone working on actual open source models (ie. training data provided):
"The most successful open models will be complementary tools to closed agents. This is a path for open models to complement and accelerate the frontier of progress."
... "These models need to be almost brain-numbingly boring and specific. In a world dominated by coding agents, I want to build open models that Claude Code is desperate to use as a tool, letting its sub agents unlock entirely new areas of work. This is possible, but remarkably under-explored. Small models from the likes of Qwen and co. are still marketed on general-task benchmarks. The hype of “open models catching the frontier” distracts the world from this very large area of demand."
https://www.interconnects.ai/p/the-next-phase-of-open-models
Alibaba's flagship Qwen team sees a key departure right after the open-source launch of Qwen 3.5. The move raises questions about the future of agentic inflection, AI workers and enterprise-grade LLMs. What does this mean for open-source AI? Dive into the details. #QwenAI #AlibabaAI #OpenSourceLLM #EnterpriseAI
🔗 https://aidailypost.com/news/alibaba-sees-key-qwen-ai-staff-exit-after-qwen35-open-source-release
Alibaba just released the Qwen‑3.5‑Medium model as open‑source, delivering Sonnet 4.5‑level performance on a single GPU. It uses a Mixture‑of‑Experts architecture and a new “Thinking Mode” to boost AI inference efficiency while staying lightweight. Dive into the details and see how this could reshape open‑source LLM development. #Qwen3_5 #OpenSourceLLM #MixtureOfExperts #ModelEfficiency
🔗 https://aidailypost.com/news/alibaba-open-sources-qwen35-medium-models-sonnet-45-performance
RE: https://hachyderm.io/@loleg/116041701680591995
Collected impressions of the #OpenSourceLLM summit on my personal blog https://log.alets.ch/121/
Running with Swiss perseverance if not quite precision today 😅 Lots of questions and an avalanche of content shared in the past 4 hours.
The #Qwen team can’t be very specific on the compute capabilities, but they seem ready for the road ahead:
• Hybrid Architecture works
• Multimodal native pretraining + post-training
• Coding with an LLM with vision
• Tackle long-horizon agentic tasks