Timezone: »
Graph neural networks (GNNs) have been demonstrated to be an effective model for learning tasks related to graph structured data. Different from classical deep neural networks which handle relatively small individual samples, GNNs process very large graphs, which must be partitioned and processed in a distributed manner. We present Roc, a distributed multi-GPU framework for fast GNN training and inference on graphs. Roc is up to 4.6x faster than existing GNN frameworks on a single machine, and can scale to multiple GPUs on multiple machines. This performance gain is mainly enabled by Roc's graph partitioning and memory management optimizations. Besides performance acceleration, the better scalability of Roc also enables the exploration of more sophisticated GNN architectures on large, real-world graphs. We demonstrate that a class of GNN architectures significantly deeper and larger than the typical two-layer models can achieve new state-of-the-art classification accuracy on the widely used Reddit dataset.
Author Information
Zhihao Jia (Stanford University)
Sina Lin (Microsoft)
Mingyu Gao (Tsinghua University)
Matei Zaharia (Stanford and Databricks)

Matei Zaharia is an Associate Professor of Computer Science at Stanford (moving to UC Berkeley later this year) and Chief Technologist and Cofounder of Databricks. His research has spanned distributed systems, databases, security and machine learning, with the most recent focus on systems for machine learning, natural language processing, and information retrieval. Matei started and contributed to multiple widely used open source projects including Apache Spark (his PhD project at UC Berkeley), MLflow, Dolly, Delta Lake, and ColBERT. His research was recognized through the 2014 ACM Doctoral Dissertation Award, an NSF CAREER Award, and the US Presidential Early Career Award for Scientists and Engineers (PECASE).
Alex Aiken (Stanford University)
Related Events (a corresponding poster, oral, or spotlight)
-
2020 Oral: Improving the Accuracy, Scalability, and Performance of Graph Neural Networks with Roc »
Tue. Mar 3rd 04:55 -- 05:20 PM Room Ballroom A
More from the Same Authors
-
2023 Poster: MegaBlocks: Efficient Sparse Training with Mixture-of-Experts »
Trevor Gale · Deepak Narayanan · Cliff Young · Matei Zaharia -
2023 Invited Talk: Improving the Quality and Factuality of Large Language Model Applications »
Matei Zaharia -
2021 Poster: IOS: Inter-Operator Scheduler for CNN Acceleration »
Yaoyao Ding · Ligeng Zhu · Zhihao Jia · Gennady Pekhimenko · Song Han -
2021 Oral: IOS: Inter-Operator Scheduler for CNN Acceleration »
Yaoyao Ding · Ligeng Zhu · Zhihao Jia · Gennady Pekhimenko · Song Han -
2020 Workshop: MLOps Systems »
Debo Dutta · Matei Zaharia · Ce Zhang -
2020 Oral: MLPerf Training Benchmark »
Peter Mattson · Christine Cheng · Gregory Diamos · Cody Coleman · Paulius Micikevicius · David Patterson · Hanlin Tang · Gu-Yeon Wei · Peter Bailis · Victor Bittorf · David Brooks · Dehao Chen · Debo Dutta · Udit Gupta · Kim Hazelwood · Andy Hock · Xinyuan Huang · Daniel Kang · David Kanter · Naveen Kumar · Jeffery Liao · Deepak Narayanan · Tayo Oguntebi · Gennady Pekhimenko · Lillian Pentecost · Vijay Janapa Reddi · Taylor Robie · Tom St John · Carole-Jean Wu · Lingjie Xu · Cliff Young · Matei Zaharia -
2020 Poster: Willump: A Statistically-Aware End-to-end Optimizer for Machine Learning Inference »
Peter Kraft · Daniel Kang · Deepak Narayanan · Shoumik Palkar · Peter Bailis · Matei Zaharia -
2020 Poster: MLPerf Training Benchmark »
Peter Mattson · Christine Cheng · Gregory Diamos · Cody Coleman · Paulius Micikevicius · David Patterson · Hanlin Tang · Gu-Yeon Wei · Peter Bailis · Victor Bittorf · David Brooks · Dehao Chen · Debo Dutta · Udit Gupta · Kim Hazelwood · Andy Hock · Xinyuan Huang · Daniel Kang · David Kanter · Naveen Kumar · Jeffery Liao · Deepak Narayanan · Tayo Oguntebi · Gennady Pekhimenko · Lillian Pentecost · Vijay Janapa Reddi · Taylor Robie · Tom St John · Carole-Jean Wu · Lingjie Xu · Cliff Young · Matei Zaharia -
2020 Poster: Model Assertions for Monitoring and Improving ML Models »
Daniel Kang · Deepti Raghavan · Peter Bailis · Matei Zaharia -
2020 Oral: Model Assertions for Monitoring and Improving ML Models »
Daniel Kang · Deepti Raghavan · Peter Bailis · Matei Zaharia -
2020 Oral: Willump: A Statistically-Aware End-to-end Optimizer for Machine Learning Inference »
Peter Kraft · Daniel Kang · Deepak Narayanan · Shoumik Palkar · Peter Bailis · Matei Zaharia