LLVM Social @ Cambridge

Date: September 3rd 2024
Time: 15:00 (Talk), 16:00-20:00 (Social)
Location: William Gates Building, 15 JJ Thomson Ave, Cambridge CB3 0FD
Rooms: LT1 (Talk), The Street (Social)
Hosts: Markus Böck, Tobias Grosser

Join us for a relaxed chat about compilers, while socializing over refreshments. Our social is open to students, academics, professional developers and really anyone interested in compilation. We welcome beginners as well as experts. Our social is an unguided space offered for you to get to know people, try out some new ideas, get feedback on your code, or pair-program on a difficult program. Come with just a paper notebook or bring your laptop to hack on some in-progress patches.

This social is traditionally organized by the LLVM community, but is open to all (potential) compiler enthusiasts.

Talk 1: Quidditch: An end-to-end deep learning compiler for highly-concurrent accelerators with software-managed caches (Markus Böck, University of Cambridge)

The wide adoption of Deep Neural Networks and the resulting desire for more hardware resources has fueled the rapid development of innovative custom hardware accelerators that are increasingly difficult to program.
Many proposed hardware designs are only evaluated with hand-written micro-kernels, and the few evaluated on entire neural networks typically require significant investments in building the necessary software stacks.
Highly sophisticated neural network compilers emerged to generate DNNs out of expert-written microkernels, but they were traditionally hand-crafted for each platform, which prevented both scaling and synergy with industry-supported compilation flows.    
We present Quidditch, a novel neural network compiler and runtime, that provides an end-to-end workflow from a high-level network description to high-performance code running on ETH Occamy, one of the first chiplet-based AI research hardware accelerators.
Quidditch builds on IREE, an industry-strength AI compiler and runtime focused on GPUs. Quidditch imports NNs from PyTorch, JAX, and Tensorflow and offers optimisations such as fusion, scheduling, buffer allocation, memory and multi-level concurrency-guided tiling and asynchronous memory transfers to scratchpads.
We present a set of preliminary novel optimisations, SSA-based double-buffering and barrier management for scratchpads, and redundant transfer elimination tailored for explicitly managed memory. We pair this with a high-performance microkernel generator, which enables us to run full DNNs with full FPU occupancy and a more than 20x speed-up over IREE’s generic LLVM backend on our custom hardware accelerator.
By providing key building blocks for scaling AI accelerator compilation to full neural networks, we aim to accelerate the evaluation of custom AI hardware and, as a result, AI hardware development overall.

Talk 2: Mojo's Wishlist for MLIR 2.0 (Jeff Niu, Modular)

Mojo is a systems programming language built natively on top of MLIR and leverages MLIR to build state-of-the-art compiler technology. Mojo is the foundation of Modular's heterogeneous compute platform, enabling performance portability across different hardware and application domains.

After 2 years of building Mojo with MLIR, design misalignments between the compiler infrastructure and the desired language semantics have clearly emerged. This talk will delve into what an ideal MLIR 2.0 would look like purely in the context of the design of Mojo: first-class dependent types, unified types and attributes, control flow, etc. We will also explore our challenges scaling MLIR compilation to the massive amounts of code backing LLMs and our experience building a multithreaded compiler.
Sign in to Google to save your progress. Learn more
Full Name
*
Email *
Affiliation
*
Can we share your Name, E-Mail and Affiliation with other attendants? *
Attendance
*
Please note which parts of the event you want to participate in
Required
Please note that when you attend this event, you enter an area where photography, audio, and video recording may occur. By entering the attending, you consent to such recording media and its release, publication, exhibition or reproduction.
*
Dietary Requirements
*
Required
Submit
Clear form
Never submit passwords through Google Forms.
This content is neither created nor endorsed by Google. - Terms of Service - Privacy Policy

Does this form look suspicious? Report