Black Forest Labs' new Self‑Flow framework cuts multimodal AI training time by 2.8× versus REPA, thanks to smarter feature alignment and better computational efficiency. Open‑source researchers can now train larger models faster. Dive into the details to see how this could reshape your ML pipelines. #SelfFlow #MultimodalAI #AITraining #ComputationalEfficiency

🔗 https://aidailypost.com/news/black-forest-labs-self-flow-speeds-multimodal-ai-training-28-faster

Nghiên cứu mới giới thiệu DCL-ENAS, cải tiến cho Evolutionary Neural Architecture Search (ENAS) để tự động thiết kế mạng nơ-ron hiệu quả hơn. DCL-ENAS dùng học tương phản kép, giúp đào tạo bộ dự đoán với chi phí thấp hơn và độ chính xác cao hơn, bằng cách dự đoán hiệu suất tương đối của các kiến trúc. Phương pháp này đạt độ chính xác cao nhất trên NASBench-101/201 và cải thiện phân loại nhịp tim ECG đáng kể.

#AI #MachineLearning #DeepLearning #NAS #ENAS #DCLENAS #ComputationalEfficiency
#TríTuệ

DeepSeek tests “sparse attention” to slash AI processing costs

Chinese lab’s v3.2 release explores a technique that could make running AI far less costly.

Ars Technica
🚀 Behold the new epoch of image generation: an epic tome on how to press Ctrl+C and Ctrl+V to save precious computation cycles. Surely, the pinnacle of human ingenuity was reached by simply learning to reuse things. 😜 Because who needs #innovation when you've mastered the art of the copy-paste? 🖨️
https://arxiv.org/abs/2508.21032 #imagegeneration #copyandpaste #techhumor #computationalefficiency #HackerNews #ngated
Reusing Computation in Text-to-Image Diffusion for Efficient Generation of Image Sets

Text-to-image diffusion models enable high-quality image generation but are computationally expensive. While prior work optimizes per-inference efficiency, we explore an orthogonal approach: reducing redundancy across correlated prompts. Our method leverages the coarse-to-fine nature of diffusion models, where early denoising steps capture shared structures among similar prompts. We propose a training-free approach that clusters prompts based on semantic similarity and shares computation in early diffusion steps. Experiments show that for models trained conditioned on image embeddings, our approach significantly reduces compute cost while improving image quality. By leveraging UnClip's text-to-image prior, we enhance diffusion step allocation for greater efficiency. Our method seamlessly integrates with existing pipelines, scales with prompt sets, and reduces the environmental and financial burden of large-scale text-to-image generation. Project page: https://ddecatur.github.io/hierarchical-diffusion/

arXiv.org