Timezone: »
Binarized neural networks have attracted much recent attention due to their promise of making convolutional neural networks fast and compact. However, these benefits have proven hard to realize in practice. In this paper, we identify the underlying barriers to high performance and propose solutions from missing implementations for certain operations to carefully scheduled library support for binarized linear algebra operations. The combination of these innovations allows us to report the first measured end-to-end speedups for binarized networks. For instance, we show a 6.3_ speedup over a standard VGGNet variant at state-of-the-art (64.2% for top-1 binarized classification of ImageNet) accuracy. More broadly speedups range from 4-12_ and the techniques we propose are crucial to achieving them.
Author Information
Joshua Fromm (University of Washington)
Meghan Cowan (University of Washington)
Matthai Philipose (Microsoft Research)
Luis Ceze (University of Washington and OctoML)
Shwetak Patel (University of Washington)
Related Events (a corresponding poster, oral, or spotlight)
-
2020 Poster: Riptide: Fast End-to-End Binarized Neural Networks »
Tue. Mar 3rd 12:30 -- 03:00 AM Room Ballroom A #29
More from the Same Authors
-
2022 Poster: DietCode: Automatic Optimization for Dynamic Tensor Programs »
Bojian Zheng · Ziheng Jiang · Cody Hao Yu · Haichen Shen · Joshua Fromm · Yizhi Liu · Yida Wang · Luis Ceze · Tianqi Chen · Gennady Pekhimenko -
2022 Poster: SRIFTY: Swift and Thrifty Distributed Neural Network Training on the Cloud »
Liang Luo · Peter West · Pratyush Patel · Arvind Krishnamurthy · Luis Ceze -
2022 Oral: SRIFTY: Swift and Thrifty Distributed Neural Network Training on the Cloud »
Liang Luo · Liang Luo · Peter West · Peter West · Pratyush Patel · Pratyush Patel · Arvind Krishnamurthy · Luis Ceze · Luis Ceze -
2022 Oral: DietCode: Automatic Optimization for Dynamic Tensor Programs »
Bojian Zheng · Ziheng Jiang · Cody Hao Yu · Haichen Shen · Joshua Fromm · Yizhi Liu · Yida Wang · Luis Ceze · Tianqi Chen · Gennady Pekhimenko -
2021 : Thoughts on Research, Community and Impact »
Luis Ceze -
2021 : Panel Discussion »
Luis Ceze · Cliff Young · Chris Lattner -
2020 Poster: PLink: Discovering and Exploiting Locality for Accelerated Distributed Training on the public Cloud »
Liang Luo · Peter West · Jacob Nelson · Arvind Krishnamurthy · Luis Ceze -
2020 Oral: PLink: Discovering and Exploiting Locality for Accelerated Distributed Training on the public Cloud »
Liang Luo · Peter West · Jacob Nelson · Arvind Krishnamurthy · Luis Ceze