Getting Started
Schedule
Main Conference
Invited Talks
Papers
Awards
Round Table Discussion
Workshops
Sponsors
Login
Show Detail »
Schedule
Sun
Tue
Wed
Thu
Timezone:
America/Los_Angeles
Filter Events:
Break
Invited Talk
Poster
Registration Desk
Remarks
Round Table Discussion
Workshop
Filter Rooms:
Ballroom B
Ballroom C
Room 202
Room 234
Room 238
Room 239
Room 240/241
Room 242/243
Room 246/247
Room 248/249
SUN 4 JUN
11 a.m.
Registration Desk:
Registration Desk
(ends 2:00 PM)
MON 5 JUN
4:30 a.m.
Registration Desk:
Registration Desk
(ends 1:00 PM)
5 a.m.
Coffee Break
5:45 a.m.
Opening Remarks:
Opening Remarks
(ends 6:00 AM)
6 a.m.
Parallel and Distributed Systems 1: Parallelism
[6:00-7:00]
Efficient Training of Large Language Models Using Pipelining and Fisher Information Matrices
Tutel: Adaptive Mixture-of-Experts at Scale
Breadth-First Pipeline Parallelism
(ends 7:00 AM)
7 a.m.
Coffee Break
7:30 a.m.
Invited Talk:
Improving the Quality and Factuality of Large Language Model Applications
Matei Zaharia
(ends 8:30 AM)
8:30 a.m.
Lunch Break, on your own
9:45 a.m.
Round Table Discussion:
Round Table Discussion
(ends 10:30 AM)
10:30 a.m.
Memory Optimization
[10:30-11:50]
Reducing Activation Recomputation in Large Transformer Models
RevBiFPN: The Fully Reversible Bidirectional Feature Pyramid Network
Transcending Runtime-Memory Tradeoffs in Checkpointing by being Fusion Aware
Safe Optimized Static Memory Allocation for Parallel Deep Learning
(ends 11:50 AM)
11:50 a.m.
Coffee Break
12:20 p.m.
Correctness and Security
[12:20-1:40]
Validating Large Language Models with ReLM
SysNoise: Exploring and Benchmarking Training-Deployment System Inconsistency
Be Careful with PyPI Packages: You May Unconsciously Spread Backdoor Model Weights
Building Verified Neural Networks for Computer Systems with Ouroboros
(ends 1:40 PM)
1:40 p.m.
Sparsity 1: Models and Algorithms
[1:40-2:40]
MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Uniform Sparsity in Deep Neural Networks
Cuttlefish: Low-rank Model Training without All The Tuning
(ends 2:40 PM)
2:40 p.m.
Poster Session/Reception
[2:40-5:00]
(ends 5:00 PM)
TUE 6 JUN
4:30 a.m.
Registration Desk:
Registration Desk
(ends 1:00 PM)
5:30 a.m.
Coffee Break
6 a.m.
Measurement and Analysis
[6:00-7:00]
Efficiently Scaling Transformer Inference
Hotline Profiler: Automatic Annotation and A Multi-Scale Timeline for Visualizing Time-Use in DNN Training
ApproxCaliper: A Programmable Framework for Application-aware Neural Network Optimization
(ends 7:00 AM)
7 a.m.
Coffee Break
7:30 a.m.
Invited Talk:
Do we need Attention?
Alexander Rush
(ends 8:30 AM)
8:30 a.m.
Lunch Break, on your own
10:30 a.m.
Parallel and Distributed Systems 2: Communication
[10:30-11:50]
Cupcake: A Compression Scheduler for Scalable Communication-Efficient Distributed Training
Communication-Efficient Graph Neural Networks with Probabilistic Neighborhood Expansion Analysis and Caching
Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training
On Optimizing the Communication of Model Parallelism
(ends 11:50 AM)
11:50 a.m.
Coffee Break
12:20 p.m.
Federated Learning
[12:20-1:40]
FedTree: A Federated Learning System For Trees
FLINT: A Platform for Federated Learning Integration
On Noisy Evaluation in Federated Hyperparameter Tuning
GlueFL: Reconciling Client Sampling and Model Masking for Bandwidth Efficient Federated Learning
(ends 1:40 PM)
1:40 p.m.
ML for Systems
[1:40-3:00]
AutoScratch: ML-Optimized GPU Cache Management
GiPH: Generalizable Placement Learning for Adaptive Heterogeneous Computing
Learning to Parallelize with OpenMP by Augmented Heterogeneous AST Representation
Virtual Machine Allocation with Lifetime Predictions
(ends 3:00 PM)
WED 7 JUN
4:30 a.m.
Registration Desk:
Registration Desk
(ends 1:00 PM)
5:30 a.m.
Coffee Break
6 a.m.
Compilers
[6:00-7:00]
ALCOP: Automatic Load-Compute Pipelining in Deep Learning Compiler for AI-GPUs
Sirius: Harvesting Whole-Program Optimization Opportunities for DNNs
X-RLflow: Graph Reinforcement Learning for Neural Network Subgraph Transformation
(ends 7:00 AM)
7 a.m.
Coffee Break
7:30 a.m.
Emerging Models and Domains
[7:30-8:30]
XRBench: An Extended Reality (XR) Machine Learning Benchmark Suite for the Metaverse
Renee: END-TO-END TRAINING OF EXTREME CLASSIFICATION MODELS
HyperGef: A Framework Enabling Efficient Fusion for Hypergraph Neural Network on GPUs
(ends 8:30 AM)
8:30 a.m.
Lunch Break, on your own
10:30 a.m.
Sparsity 2: Systems
[10:30-12:20]
Efficient GPU Kernels for N:M-Sparse Weights in Deep Learning
Unified Convolution Framework: A compiler-based approach to support sparse convolutions
Exploiting Hardware Utilization and Adaptive Dataflow for Efficient Sparse Convolution in 3D Point Clouds
Sparsity-Aware Memory Interface Architecture using Stacked XORNet Compression for Accelerating Pruned-DNN Models
(ends 12:20 PM)
11:50 a.m.
Coffee Break
12:20 p.m.
Storage, Scheduling, and Networking
[12:20-1:40]
Pre-trained Neural Cost Models for Efficient Embedding Table Sharding in Deep Learning Recommendation Models
μ-TWO: MULTI-MODEL TRAINING WITH ORCHESTRATION AND MEMORY OPTIMIZATION
DISTRIBUTED DEEP LEARNING BUILT ON TENSOR-OPTIMIZED REMOTE PROCEDURE CALLS
RecD: Deduplication for End-to-End Deep Learning Recommendation Model Training Infrastructure
(ends 1:40 PM)
1:40 p.m.
Edge
[1:40-2:40]
Practical Edge Kernels for Integer-Only Vision Transformers Under Post-training Quantization
Edge Impulse: An MLOps Platform for Tiny Machine Learning
SUBGRAPH STATIONARY HARDWARE-SOFTWARE INFERENCE CO-DESIGN
(ends 2:40 PM)
THU 8 JUN
4 a.m.
Registration Desk:
Registration Desk
(ends 9:00 AM)
5 a.m.
Workshop:
Workshop on Systems for Next-Gen AI Paradigms
(ends 2:00 PM)
Workshop:
2nd Workshop on Practical Adoption Challenges of ML for Systems in Industry
(ends 2:00 PM)
Workshop:
Benchmarking Machine Learning Workloads on Emerging Hardware
(ends 2:00 PM)
5:30 a.m.
Workshop:
Research On Algorithms & Data Structures (ROADS) to Mega-AI Models
(ends 2:00 PM)
5:45 a.m.
Workshop:
Workshop on Decentralized and Collaborative Learning
(ends 3:20 PM)
5:50 a.m.
Workshop:
Workshop on Federated Learning Systems
(ends 2:05 PM)
6 a.m.
Workshop:
Resource-Constrained Learning in Wireless Networks
(ends 12:00 PM)
Workshop:
The 3rd On-Device Intelligence Workshop
(ends 2:15 PM)
7 a.m.
Coffee Break
noon
Coffee Break
MLSys uses cookies to remember that you are logged in. By using our websites, you agree to the placement of these cookies.
Our Privacy Policy »
Accept Cookies