Latest Papers: MLIR, Sparse Data, Tensor Formats (Oct 2025)

by SLV Team 60 views

Hey everyone! Check out the latest research papers in MLIR, Sparse Data Structures, and Tensor Formats. For a better reading experience and more papers, make sure to visit the Github page.

MLIR

Title Date Comment
Interleaved Learning and Exploration: A Self-Adaptive Fuzz Testing Framework for MLIR 2025-10-09
Mojo: MLIR-Based Performance-Portable HPC Science Kernels on GPUs for the Python Ecosystem 2025-09-25
Accepted at the IEEE/ACM SC25 Conference WACCPD Workshop.

Accepted at the IEEE/ACM SC25 Conference WACCPD Workshop. The International Conference for High Performance Computing, Networking, Storage, and Analysis, St. Louis, MO, Nov 16-21, 2025. 15 pages, 7 figures. WFG and TM contributed equally

WAMI: Compilation to WebAssembly through MLIR without Losing Abstraction 2025-06-19
DESIL: Detecting Silent Bugs in MLIR Compiler Infrastructure 2025-04-02
Building Bridges: Julia as an MLIR Frontend 2025-02-14
This is the extended abstract of a master's thesis

This is the extended abstract of a master's thesis, hosted at https://lib.ugent.be/en/catalog/rug01:003212846?i=0

Supervised by: Prof. Bjorn De Sutter with counselling from: Dr. Tim Besard and Thomas Faingnaert

Fully integrating the Flang Fortran compiler with standard MLIR 2024-09-27
Author accepted version

Author accepted version, to appear in proceedings of the tenth annual workshop on the LLVM compiler infrastructure in HPC

A Reinforcement Learning Environment for Automatic Code Optimization in the MLIR Compiler 2024-09-17
The MLIR Transform Dialect. Your compiler is more powerful than you think 2024-09-09
Fuzzing MLIR Compilers with Custom Mutation Synthesis 2024-08-27
DSP-MLIR: A MLIR Dialect for Digital Signal Processing 2024-08-20
Towards a high-performance AI compiler with upstream MLIR 2024-04-15
13 pages, 8 figures

13 pages, 8 figures, presented at CGO C4ML 2024 & MLIR Workshop EuroLLVM 2024

An Optimizing Framework on MLIR for Efficient FPGA-based Accelerator Generation 2024-01-10 Accepted by HPCA2024
Experiences Building an MLIR-based SYCL Compiler 2023-12-20
12 pages, 3 figures

12 pages, 3 figures To be published in International Symposium on Code Generation and Optimization (CGO) 2024

Fortran performance optimisation and auto-parallelisation by leveraging MLIR-based domain specific abstractions in Flang 2023-10-03
Author accepted version of paper

Author accepted version of paper in ACM Workshops of The International Conference on High Performance Computing, Network, Storage, and Analysis (SC-W 2023)

Platform-Aware FPGA System Architecture Generation based on MLIR 2023-09-22
Accepted for presentation

Accepted for presentation at the CPS workshop 2023 (http://www.cpsschool.eu/cps-workshop)


Diving Deep into MLIR Research Papers

In the realm of compiler technology, MLIR (Multi-Level Intermediate Representation) is quickly becoming a game-changer. These latest papers in MLIR highlight the cutting-edge research and developments in this field. From fuzz testing frameworks to domain-specific abstractions, MLIR's versatility is truly remarkable. One standout paper, "Interleaved Learning and Exploration: A Self-Adaptive Fuzz Testing Framework for MLIR," showcases the importance of robust testing methodologies in compiler development. Fuzzing, the process of feeding invalid or unexpected inputs to a system, is a critical technique for identifying bugs and vulnerabilities. This paper introduces a novel approach to fuzzing MLIR compilers, ensuring that they are reliable and resilient. Another noteworthy paper, "Mojo: MLIR-Based Performance-Portable HPC Science Kernels on GPUs for the Python Ecosystem," explores the potential of MLIR in high-performance computing (HPC). By leveraging MLIR, researchers have developed a system that allows for the efficient execution of scientific kernels on GPUs, making it easier for Python developers to tap into the power of parallel processing. Furthermore, the paper "Building Bridges: Julia as an MLIR Frontend" investigates the integration of the Julia programming language with MLIR. Julia, known for its speed and expressiveness, can benefit greatly from the optimizations that MLIR provides. This integration opens up exciting possibilities for numerical computing and scientific simulations. The range of topics covered in these papers demonstrates the breadth and depth of MLIR research. Whether it's detecting silent bugs, optimizing code, or generating hardware accelerators, MLIR is proving to be a powerful tool for compiler developers and researchers alike. It's an exciting time to be involved in MLIR, and these papers offer a glimpse into the future of compiler technology.

Sparse Data Structure

Title Date Comment
A Novel Compiler Transformation for Fast Sparse Matrix Multiplication in GPUs 2025-06-18
Interface for Sparse Linear Algebra Operations 2024-11-20 43 pages
Architecture Specific Generation of Large Scale Lattice Boltzmann Methods for Sparse Complex Geometries 2024-08-13 16 pages, 19 figures
Scorch: A Library for Sparse Deep Learning 2024-06-20 25 pages, 8 figures
Filtering After Shading With Stochastic Texture Filtering 2024-05-14
arXiv admin note

arXiv admin note: substantial text overlap with arXiv:2305.05810

Tascade: Hardware Support for Atomic-free, Asynchronous and Efficient Reduction Trees 2024-04-22
TCAM-SSD: A Framework for Search-Based Computing in Solid-State Drives 2024-03-11
UniSparse: An Intermediate Language for General Sparse Format Customization 2024-03-09
to be published in OOPSLA'24

to be published in OOPSLA'24

SUREL+: Moving from Walks to Sets for Scalable Subgraph-based Graph Representation Learning 2023-12-27
This is an extended version of the full paper

This is an extended version of the full paper that appeared in PVLDB 16.11(VLDB 2023)

Contracting Tsetlin Machine with Absorbing Automata 2023-10-17
Accepted to ISTM2023

Accepted to ISTM2023. 7 pages, 8 figures

Sparse Stream Semantic Registers: A Lightweight ISA Extension Accelerating General Sparse Linear Algebra 2023-10-02
15 pages, 8 figures

15 pages, 8 figures. Accepted for publication in IEEE TPDS

Massive Data-Centric Parallelism in the Chiplet Era 2023-08-11
Stochastic Texture Filtering 2023-05-15 15 pages
FSCNN: A Fast Sparse Convolution Neural Network Inference System 2022-12-17
technical report, sparse CNN

technical report, sparse CNN

SpDISTAL: Compiling Distributed Sparse Tensor Computations 2022-07-28

Exploring Sparse Data Structures: The Latest Research

Sparse data structures are vital in modern computing, especially when dealing with large datasets where most elements are zero. The efficiency of these structures can significantly impact performance in applications ranging from machine learning to scientific computing. The latest research papers on sparse data structures showcase innovative approaches to tackling the challenges posed by these data types. One fascinating paper, "A Novel Compiler Transformation for Fast Sparse Matrix Multiplication in GPUs," addresses the critical issue of optimizing sparse matrix multiplication on GPUs. Matrix multiplication is a fundamental operation in many scientific and engineering applications, and optimizing it for sparse matrices can lead to significant speedups. This paper introduces a new compiler transformation technique that leverages the parallel processing capabilities of GPUs to accelerate sparse matrix multiplication. Another notable paper, "UniSparse: An Intermediate Language for General Sparse Format Customization," presents a novel intermediate language designed to facilitate the customization of sparse formats. Different applications have different requirements for sparse data storage, and UniSparse provides a flexible framework for tailoring sparse formats to specific needs. This level of customization can lead to substantial performance improvements in various applications. Furthermore, the paper "Scorch: A Library for Sparse Deep Learning" introduces a library specifically designed for sparse deep learning. Deep learning models often involve large numbers of parameters, and many of these parameters may be zero or close to zero. By leveraging sparse data structures, Scorch aims to reduce memory consumption and improve the efficiency of deep learning computations. These papers demonstrate the ongoing efforts to enhance the performance and flexibility of sparse data structures. As data continues to grow in size and complexity, the importance of efficient sparse data representations will only increase. The research highlighted here offers valuable insights into the future of sparse data processing.

Tensor Formats

Title Date Comment
Inexact subspace projection methods for low-rank tensor eigenvalue problems 2025-10-15 25 pages, 7 figures
Model order reduction of parametric dynamical systems by slice sampling tensor completion 2025-10-11
A new cross approximation for Tucker tensors and its application in Tucker-Anderson Acceleration 2025-09-23
ReLATE: Learning Efficient Sparse Encoding for High-Performance Tensor Decomposition 2025-08-29
Binsparse: A Specification for Cross-Platform Storage of Sparse Matrices and Tensors 2025-06-23
Wildfire Detection Using Vision Transformer with the Wildfire Dataset 2025-05-23
Published at ASEE NE 2025

Published at ASEE NE 2025

Irregular Tensor Low-Rank Representation for Hyperspectral Image Representation 2025-05-18 Accepted by TIP
Dynamical low-rank tensor approximations to high-dimensional parabolic problems: existence and convergence of spatial discretizations 2025-05-16
A vector bundle approach to Nash equilibria 2025-04-04
34 pages, 2 tables

34 pages, 2 tables. Comments are welcome!

Accelerating Sparse Tensor Decomposition Using Adaptive Linearized Representation 2025-03-15
Accepted to TPDS 2025

Accepted to TPDS 2025

A mesh-free hybrid Chebyshev-Tucker tensor format with applications to multi-particle modelling 2025-03-03
SySTeC: A Symmetric Sparse Tensor Compiler 2025-01-23
Provable Low-Rank Tensor-Train Approximations in the Inverse of Large-Scale Structured Matrices 2025-01-13
Desingularization of bounded-rank tensor sets 2024-11-21
41 pages, 10 figures, 1 table

41 pages, 10 figures, 1 table

The Continuous Tensor Abstraction: Where Indices are Real 2024-07-01

Recent Advances in Tensor Formats: A Deep Dive

Tensor formats are essential for representing and manipulating multi-dimensional arrays, which are fundamental in various fields such as machine learning, data analysis, and scientific computing. The way tensors are structured and stored can significantly impact the efficiency of computations performed on them. The latest research in tensor formats explores novel methods for optimizing tensor representations and algorithms. The paper "Inexact subspace projection methods for low-rank tensor eigenvalue problems" presents new techniques for solving tensor eigenvalue problems, which arise in many applications, including signal processing and data mining. By employing inexact subspace projection methods, the authors aim to improve the computational efficiency of eigenvalue calculations. Another intriguing paper, "ReLATE: Learning Efficient Sparse Encoding for High-Performance Tensor Decomposition," focuses on learning efficient sparse encodings for tensor decomposition. Tensor decomposition is a powerful tool for dimensionality reduction and feature extraction, but it can be computationally expensive for large tensors. This paper introduces a learning-based approach to identify sparse encodings that preserve the essential information in the tensor while reducing storage and computational costs. Furthermore, the paper "Binsparse: A Specification for Cross-Platform Storage of Sparse Matrices and Tensors" proposes a specification for cross-platform storage of sparse matrices and tensors. Standardization in storage formats is crucial for interoperability and data exchange between different systems and applications. Binsparse aims to provide a common format for representing sparse tensors, facilitating collaboration and code reuse across various platforms. These research papers highlight the diverse challenges and opportunities in the field of tensor formats. From developing efficient algorithms for tensor decomposition to designing standardized storage formats, the ongoing research efforts are pushing the boundaries of tensor computing and enabling new applications in data science and beyond. As the demand for processing large, multi-dimensional datasets continues to grow, the importance of innovative tensor formats will only increase.