ABCDEFG
1
SectionTime (EST)TitleSpeakerAttendanceWorkshopSession Chair
2
Opening9:00AM - 9:05 AMOpening and AnnouncementsAmir Yazdanbakhsh (Google)
3
Keynote Speaker9:05AM - 9:50AMML4ML: The Intriguing Interplay between Moore’s Law & Machine LearningParthasarathy Ranganathan (Google)Neeraja J. Yadwadkar (UT Austin)
4
9:50AM - 10:00AMBreak
5
Session 110:00AM - 10:10AMLightweight ML-based Runtime Prefetcher Selection on Many-core PlatformsErika S AlcortaIn-PersonMLArchSysAmir Yazdanbakhsh (Google DeepMind)
6
10:10AM - 10:20AMDOSA: One-Loop DSE for DNN Accelerators Using Differentiable ModelsCharles HongIn-PersonMLArchSys
7
10:20AM - 10:30AMTowards Efficient Multi-Agent Learning SystemsKailash GogineniIn-PersonMLArchSys
8
10:30AM -10:40AMParaGAN: A Cloud Training Framework for Generative Adversarial NetworksZiji ShiIn-PersonMLArchSys
9
10:40AM - 10:50AMSample-Efficient Mapspace Optimization for DNN Accelerators with Bayesian LearningGrace DinhIn-PersonMLArchSys
10
10:50AM - 11:00AMOnline Learning for Right-Sizing Serverless FunctionsPrasoon SinhaIn-PersonMLArchSys
11
11:00AM - 11:05AMBuffer
12
11:05PM - 11:20PMCoffee Break
13
Keynote Speaker11:20AM - 12:05PMLLM Training at Wafer-ScaleValavan Manohararajah (Cerebras)Bahar Asgari (UMD)
14
Session 212:05PM - 12:15PMScaling Infrastructure to Support Multi-Trillion Parameter LLM TrainingMikhail IsaevIn-PersonASSYST
15
12:15PM - 12:25PMAccuracy Boosters: Epoch-Driven Mixed-Mantissa Block Floating Point for DNN TrainingSimla Burcu HarmaIn-PersonMLArchSys
16
12:25PM - 12:30PMBufferBuffer
17
12:30PM - 1:40PMLunch
18
Keynote Speaker1:40PM - 2:25PMFaster Neural Network Training, AlgorithmicallyJonathan Frankle (MosaicML)Tushar Krishna (GaTech)
19
Session 32:25PM - 2:35PMAccelerating Attention Based Models via HW-SW Co-Design using Fine-Grained SparsificationAbhimanyu Rajesh BambhaniyaIn-PersonASSYST
20
2:35PM - 2:45PMEfficient Deployment of Transformer Models on Edge TPU Accelerators: A Real System EvaluationMohammadreza MohammadiIn-PersonASSYST
21
2:45PM - 2:55PMTAP: Efficient Derivation of Tensor Parallel Plans for Large Neural NetworksZiji ShiIn-PersonASSYST
22
2:55PM - 3:05PMFull Stack Optimization of Transformer InferenceColeman HooperIn-PersonASSYST
23
3:05PM - 3:15PMTowards A Reconfigurable Systolic Array with Multi-Level Packing for TransformersTiandong ZhaoIn-PersonASSYST
24
3:15PM - 3:25PMA Metric Driven Approach to Mixed Precision TrainingMitchelle RasquinhaIn-PersonASSYST
25
3:25PM - 3:30PMBuffer
26
3:30PM - 4:00PMCoffee Break
27
Keynote Speaker4:00PM - 4:45PMSoftware, Hardware, and Model Codesign for High-performance Transformer-based Large ModelsZongwei Zhou (Google)Suvinay Subramanian (Google)
28
Closing Remarks4:45PM - 4:50PMClosing RemarksSuvinay Subramanian (Google)
29
Poster Session4:50PM - 5:30 PMPoster Session