Latest Papers: MLIR, Sparse Data, Tensor Formats (Oct 2025)
Hey everyone! Check out the latest research papers in MLIR, Sparse Data Structures, and Tensor Formats. For a better reading experience and more papers, make sure to visit the Github page.
MLIR
Title | Date | Comment |
---|---|---|
Interleaved Learning and Exploration: A Self-Adaptive Fuzz Testing Framework for MLIR | 2025-10-09 | |
Mojo: MLIR-Based Performance-Portable HPC Science Kernels on GPUs for the Python Ecosystem | 2025-09-25 | Accepted at the IEEE/ACM SC25 Conference WACCPD Workshop.Accepted at the IEEE/ACM SC25 Conference WACCPD Workshop. The International Conference for High Performance Computing, Networking, Storage, and Analysis, St. Louis, MO, Nov 16-21, 2025. 15 pages, 7 figures. WFG and TM contributed equally |
WAMI: Compilation to WebAssembly through MLIR without Losing Abstraction | 2025-06-19 | |
DESIL: Detecting Silent Bugs in MLIR Compiler Infrastructure | 2025-04-02 | |
Building Bridges: Julia as an MLIR Frontend | 2025-02-14 | This is the extended abstract of a master's thesisThis is the extended abstract of a master's thesis, hosted at https://lib.ugent.be/en/catalog/rug01:003212846?i=0 |
Supervised by: Prof. Bjorn De Sutter with counselling from: Dr. Tim Besard and Thomas Faingnaert | ||
Fully integrating the Flang Fortran compiler with standard MLIR | 2024-09-27 | Author accepted versionAuthor accepted version, to appear in proceedings of the tenth annual workshop on the LLVM compiler infrastructure in HPC |
A Reinforcement Learning Environment for Automatic Code Optimization in the MLIR Compiler | 2024-09-17 | |
The MLIR Transform Dialect. Your compiler is more powerful than you think | 2024-09-09 | |
Fuzzing MLIR Compilers with Custom Mutation Synthesis | 2024-08-27 | |
DSP-MLIR: A MLIR Dialect for Digital Signal Processing | 2024-08-20 | |
Towards a high-performance AI compiler with upstream MLIR | 2024-04-15 | 13 pages, 8 figures13 pages, 8 figures, presented at CGO C4ML 2024 & MLIR Workshop EuroLLVM 2024 |
An Optimizing Framework on MLIR for Efficient FPGA-based Accelerator Generation | 2024-01-10 | Accepted by HPCA2024 |
Experiences Building an MLIR-based SYCL Compiler | 2023-12-20 | 12 pages, 3 figures12 pages, 3 figures To be published in International Symposium on Code Generation and Optimization (CGO) 2024 |
Fortran performance optimisation and auto-parallelisation by leveraging MLIR-based domain specific abstractions in Flang | 2023-10-03 | Author accepted version of paperAuthor accepted version of paper in ACM Workshops of The International Conference on High Performance Computing, Network, Storage, and Analysis (SC-W 2023) |
Platform-Aware FPGA System Architecture Generation based on MLIR | 2023-09-22 | Accepted for presentationAccepted for presentation at the CPS workshop 2023 (http://www.cpsschool.eu/cps-workshop) |
Diving Deep into MLIR Research Papers
In the realm of compiler technology, MLIR (Multi-Level Intermediate Representation) is quickly becoming a game-changer. These latest papers in MLIR highlight the cutting-edge research and developments in this field. From fuzz testing frameworks to domain-specific abstractions, MLIR's versatility is truly remarkable. One standout paper, "Interleaved Learning and Exploration: A Self-Adaptive Fuzz Testing Framework for MLIR," showcases the importance of robust testing methodologies in compiler development. Fuzzing, the process of feeding invalid or unexpected inputs to a system, is a critical technique for identifying bugs and vulnerabilities. This paper introduces a novel approach to fuzzing MLIR compilers, ensuring that they are reliable and resilient. Another noteworthy paper, "Mojo: MLIR-Based Performance-Portable HPC Science Kernels on GPUs for the Python Ecosystem," explores the potential of MLIR in high-performance computing (HPC). By leveraging MLIR, researchers have developed a system that allows for the efficient execution of scientific kernels on GPUs, making it easier for Python developers to tap into the power of parallel processing. Furthermore, the paper "Building Bridges: Julia as an MLIR Frontend" investigates the integration of the Julia programming language with MLIR. Julia, known for its speed and expressiveness, can benefit greatly from the optimizations that MLIR provides. This integration opens up exciting possibilities for numerical computing and scientific simulations. The range of topics covered in these papers demonstrates the breadth and depth of MLIR research. Whether it's detecting silent bugs, optimizing code, or generating hardware accelerators, MLIR is proving to be a powerful tool for compiler developers and researchers alike. It's an exciting time to be involved in MLIR, and these papers offer a glimpse into the future of compiler technology.
Sparse Data Structure
Exploring Sparse Data Structures: The Latest Research
Sparse data structures are vital in modern computing, especially when dealing with large datasets where most elements are zero. The efficiency of these structures can significantly impact performance in applications ranging from machine learning to scientific computing. The latest research papers on sparse data structures showcase innovative approaches to tackling the challenges posed by these data types. One fascinating paper, "A Novel Compiler Transformation for Fast Sparse Matrix Multiplication in GPUs," addresses the critical issue of optimizing sparse matrix multiplication on GPUs. Matrix multiplication is a fundamental operation in many scientific and engineering applications, and optimizing it for sparse matrices can lead to significant speedups. This paper introduces a new compiler transformation technique that leverages the parallel processing capabilities of GPUs to accelerate sparse matrix multiplication. Another notable paper, "UniSparse: An Intermediate Language for General Sparse Format Customization," presents a novel intermediate language designed to facilitate the customization of sparse formats. Different applications have different requirements for sparse data storage, and UniSparse provides a flexible framework for tailoring sparse formats to specific needs. This level of customization can lead to substantial performance improvements in various applications. Furthermore, the paper "Scorch: A Library for Sparse Deep Learning" introduces a library specifically designed for sparse deep learning. Deep learning models often involve large numbers of parameters, and many of these parameters may be zero or close to zero. By leveraging sparse data structures, Scorch aims to reduce memory consumption and improve the efficiency of deep learning computations. These papers demonstrate the ongoing efforts to enhance the performance and flexibility of sparse data structures. As data continues to grow in size and complexity, the importance of efficient sparse data representations will only increase. The research highlighted here offers valuable insights into the future of sparse data processing.
Tensor Formats
Recent Advances in Tensor Formats: A Deep Dive
Tensor formats are essential for representing and manipulating multi-dimensional arrays, which are fundamental in various fields such as machine learning, data analysis, and scientific computing. The way tensors are structured and stored can significantly impact the efficiency of computations performed on them. The latest research in tensor formats explores novel methods for optimizing tensor representations and algorithms. The paper "Inexact subspace projection methods for low-rank tensor eigenvalue problems" presents new techniques for solving tensor eigenvalue problems, which arise in many applications, including signal processing and data mining. By employing inexact subspace projection methods, the authors aim to improve the computational efficiency of eigenvalue calculations. Another intriguing paper, "ReLATE: Learning Efficient Sparse Encoding for High-Performance Tensor Decomposition," focuses on learning efficient sparse encodings for tensor decomposition. Tensor decomposition is a powerful tool for dimensionality reduction and feature extraction, but it can be computationally expensive for large tensors. This paper introduces a learning-based approach to identify sparse encodings that preserve the essential information in the tensor while reducing storage and computational costs. Furthermore, the paper "Binsparse: A Specification for Cross-Platform Storage of Sparse Matrices and Tensors" proposes a specification for cross-platform storage of sparse matrices and tensors. Standardization in storage formats is crucial for interoperability and data exchange between different systems and applications. Binsparse aims to provide a common format for representing sparse tensors, facilitating collaboration and code reuse across various platforms. These research papers highlight the diverse challenges and opportunities in the field of tensor formats. From developing efficient algorithms for tensor decomposition to designing standardized storage formats, the ongoing research efforts are pushing the boundaries of tensor computing and enabling new applications in data science and beyond. As the demand for processing large, multi-dimensional datasets continues to grow, the importance of innovative tensor formats will only increase.