PLDI 2025
Mon 16 - Fri 20 June 2025 Seoul, South Korea
Mon 16 Jun 2025 10:30 - 10:50 at Cosmos - Session 2 Chair(s): Willow Ahrens

As model sizes grow, deep neural networks that operate on different forms of sparse data, ranging from graphs to regularly sparse attention matrices, are becoming increasingly popular. However, efficiently performing training and inference on such models has been challenging. Compared to optimizing isolated sparse primitives, such models have many interacting components. As a result, optimizations and abstractions should be thought of holistically to achieve high efficiency and programming productivity. For example, when optimizing for graph machine learning, one has to reason about interactions between sparse, dense, and other components such as temporal information. In this talk, I will dive into such optimizations and abstractions we built to increase the productivity and performance of both static and temporal graph neural networks, sparse convolutional networks, and sparse attention mechanisms, achieving superior performance benefits while increasing end-user productivity.

Charith Mendis is an Assistant Professor at the University of Illinois at Urbana-Champaign. His research interests are in automating compiler construction using both formal methods and ML techniques and in building high-performance ML systems. He received his Ph.D. and Master’s from the Massachusetts Institute of Technology and his B.Sc. from the University of Moratuwa. He recently co-led the DARPA ISAT study on “ML Optimized Compilers for Heterogeneous Architectures (MOCHA).” He is the recipient of a DARPA Young Faculty Award, an NSF CAREER Award, an IEEE Micro Top Picks honorable mention, the William A. Martin outstanding master’s thesis award at MIT, a best student paper award, a best paper award, and the university gold medal for his B.Sc.

Mon 16 Jun

Displayed time zone: Seoul change

10:30 - 12:00
Session 2Sparse at Cosmos
Chair(s): Willow Ahrens Massachusetts Institute of Technology
10:30
20m
Talk
Optimizations and abstractions for sparse machine learningRecorded
Sparse
Charith Mendis University of Illinois at Urbana-Champaign
10:50
20m
Talk
Distributed Sparse Computing with Legate Sparse
Sparse
Rohan Yadav Stanford University
11:10
20m
Talk
Optimizing Recursive Sparse Computations
Sparse
Amir Shaikhha University of Edinburgh
11:30
20m
Talk
Panel 2
Sparse
Charith Mendis University of Illinois at Urbana-Champaign, Rohan Yadav Stanford University, Amir Shaikhha University of Edinburgh