🌘 透過線性代數感知編譯器實現高效稀疏計算(技術報告)
➤ 以 MLIR 驅動新一代高效能稀疏運算架構
https://www.osti.gov/biblio/3013883
本報告介紹了 LAPIS 編譯器框架的研發成果。該框架基於多層次中間表示(MLIR)構建,旨在解決稀疏線性代數運算中的效能瓶頸,並確保程式碼在多種計算架構間的移植性。透過引入創新的「Kokkos 方言」,LAPIS 成功簡化了從高階語言向底層硬體轉換的過程,並支援將 MLIR 程式碼轉換為 C++ Kokkos 程式碼,從而促進科學機器學習(SciML)模型的整合。此外,LAPIS 透過新增「分區方言」來處理分散式記憶體架構,優化了稀疏張量的分佈與通訊模式。這一框架不僅提升了稀疏與稠密矩陣運算在 GPU 上的執行效率,還廣泛應用於 GraphBLAS 資料庫(TenSQL)及複雜圖演算法,實現了開發效率與運算效能的完美平衡。
+ 終於看到有基於 MLIR 的框架能處理分散式稀疏張量了,這對於科學計算領域的大規模圖分析應用來說
#高效能計算 (HPC) #編譯器架構 #稀疏線性代數 #MLIR #分散式運算
Enabling Efficient Sparse Computations using Linear Algebra Aware Compilers (Technical Report) | OSTI.GOV

This project developed the LAPIS compiler framework, built on the Multilevel Intermediate Representation (MLIR), to optimize sparse linear algebra operations and support performance portability across diverse architectures. The main innovation of LAPIS is the Kokkos dialect, which allows for lowering codes from a high productivity language to different architectures in an elegant way. The dialect also allows the conversion of lower-level MLIR code to C++ Kokkos code, facilitating the integration of scientific machine learning (SciML) models into applications. To extend LAPIS for distributed memory architectures, a new partition dialect was created to manage the distribution of sparse tensors and express communication patterns for sparse linear algebra operations. This dialect also supports the distributed execution of operators and includes algorithmic optimizations to minimize communication to improve performance. The project also demonstrates that MLIR can enable effective linear algebra-level optimizations, improving performance on different GPUs for both sparse and dense linear algebra kernels. Key applications of LAPIS include sparse linear algebra and graph kernels, TenSQL, a relational database management solution built on GraphBLAS, and the development of subgraph isomorphism and monomorphism kernels, showcasing performance portability. In summary, the LAPIS framework supports productivity, performance, portability, and distributed memory execution, while also enabling linear algebra-level optimizations that are challenging in traditional programming languages, with successful applications ranging from simple sparse linear algebra to complex graph kernels. | OSTI.GOV

Just created our first #MLIR plugin for #cleflang and while it's within a narrow focus it has wide implications: MLKit style flat closures that can be lowered to standard LLVM. #fsharp #fstar #ocaml #dotnet

https://github.com/FidelityFramework/mlir-plugins

🌗 NVIDIA/cuda-tile:針對 CUDA 核心優化且基於 MLIR 的分塊式中間表示法與編譯器基礎設施
➤ 深度硬體協調:分塊運算優化如何重塑 GPU 效能巔峯
https://github.com/NVIDIA/cuda-tile
NVIDIA 宣佈推出開源專案「CUDA Tile IR」,這是一套基於 MLIR(多級中間表示法)架構的編譯器基礎設施,專為提升 CUDA 核心(Kernel)的執行效率而生。該技術的核心在於優化「分塊式運算」(Tile-based computation)模式,特別針對 NVIDIA 旗下的 Tensor Core 硬體單元進行深度調優。透過提供高階抽象化接口,開發者能更輕鬆地管理複雜的記憶體階層與分塊模式,從而釋放 GPU 的極致效能。此專案與最新的 CUDA Toolkit 13.1 同步釋出,象徵著 NVIDIA 在編譯器層級簡化高效能運算開發的重大突破。
+ 「NVIDIA 將 MLIR 引入 CUDA 生態系是必然的趨勢。有了官方
##GPU運算 #編譯器技術 #NVIDIA #MLIR #平行運算 #TensorCore
GitHub - NVIDIA/cuda-tile: CUDA Tile IR is an MLIR-based intermediate representation and compiler infrastructure for CUDA kernel optimization, focusing on tile-based computation patterns and optimizations targeting NVIDIA tensor core units.

CUDA Tile IR is an MLIR-based intermediate representation and compiler infrastructure for CUDA kernel optimization, focusing on tile-based computation patterns and optimizations targeting NVIDIA te...

GitHub
Normal forms for MLIR
2025 US LLVM Developers' Meeting
Alex Zinenko
https://www.youtube.com/watch?v=Esw84hH1Ed0
#LLVM #MLIR
2025 US LLVM Developers' Meeting: Normal forms for MLIR

YouTube
The #LLVM developer room is back for the 12th consecutive year at #FOSDEM, on January 31st!
We're looking for presentations on all aspects of #LLVM, #MLIR and more.
See CFP for details https://discourse.llvm.org/t/cfp-fosdem-2026-llvm-dev-room/88746.
Deadline for submissions is November 30th.
I hope to see many of you at FOSDEM!
CFP FOSDEM 2026 LLVM dev room

At FOSDEM 2026, LLVM will again participate with a dedicated devroom, on Saturday afternoon January 31st, in Brussels. As possibly the largest European Open Source Conference, FOSDEM attracts more than 600 lectures and over 8000 hackers - many core contributors of the world’s leading open source projects. Complementing the LLVM developer meetings, the devroom at FOSDEM provides a great opportunity for LLVM developers and the wider open source community to get together, connect and discuss. We ...

LLVM Discussion Forums

Interleaved Learning and Exploration: A Self-Adaptive Fuzz Testing Framework for MLIR

#MLIR #OpenCL #Testing #Package

https://hgpu.org/?p=30291

Interleaved Learning and Exploration: A Self-Adaptive Fuzz Testing Framework for MLIR

MLIR (Multi-Level Intermediate Representation) has rapidly become a foundational technology for modern compiler frameworks, enabling extensibility across diverse domains. However, ensuring the corr…

hgpu.org

Doka Studio with MLiR - Mario Marini @ Doka - 30 Aug feat. MLiR, Mario Marini

#SESH #MLiR #MarioMarini

https://sesh.sx/events/12220323

Thesis: Using Deep Reinforcement Learning for Automatic Code Optimization in the MLIR Compiler

#Performance #Physics #QCD #MLIR

https://hgpu.org/?p=30054

Using Deep Reinforcement Learning for Automatic Code Optimization in the MLIR Compiler

This work focuses on the use of deep reinforcement learning (DRL) to automate code optimization within modern compiler infrastructures. Code optimization is a critical step in program transformatio…

hgpu.org
2025 AsiaLLVM Developers' Meeting

YouTube