Tutorial
Training-Free Approaches for Edge AI: Challenges, Opportunities and Progress
Radu Marculescu · Ming Lin · Atlas Wang · Kartikeya Bhardwaj
Room 204
With the explosion in Big Data, it is often forgotten that much of the data nowadays is generated at the edge. Specifically, a major source of data is users’ endpoint devices like phones, smart watches, etc., that are connected to the internet, also known as the Internet-of-Things (IoT). Despite the huge success of deep learning (DL) in many areas (e.g., computer vision, natural language processing, etc.), the size and the computational complexity of the existing state-of-the art deep models limit the deployment of DL on resource-constrained devices and its large-scale adoption in EdgeAI. Neural architecture search (NAS) (also called AutoML) techniques have been proposed to automatically design neural architectures with reduced model sizes. The networks obtained via NAS have higher prediction accuracy and significantly fewer parameters than the hand-crafted networks. However, adapting existing NAS approaches to different hardware architectures is challenging due to their intensive computation and execution time requirements.
To address such issues, in this tutorial, we focus on the newest and perhaps the most promising breed of NAS for EdgeAI, namely approaches that are training-free and thus eminently suited for large-scale development. In particular, we plan to address a few relevant questions: What kind of system architectures can meet the AI algorithm requirements, while maximizing the prediction accuracy, inference speed as well as energy efficiency? Can we use network science or deep learning theory to understand what kind of network architectures can achieve good performance without training individual models? Can we develop efficient approaches that enable co-optimization of deep neural network accuracy and performance on real hardware platforms?
Starting from these overarching ideas, in this tutorial, we will cover both algorithmic and hardware- aware aspects of training-free model design for EdgeAI, show state-of-the-art results for relevant edge applications, and illustrate potential implications on real edge devices.
PRESENTERS: Radu Marculescu (The University of Texas at Austin), Ming Lin (Amazon), Atlas Wang (The University of Texas at Austin), Kartikeya Bhardwaj (ARM).
Conference tutorial page available here.
Schedule
Wed 1:00 p.m. - 1:15 p.m.
|
Introduction
(
Introduction
)
>
|
Radu Marculescu 🔗 |
Wed 1:15 p.m. - 1:55 p.m.
|
Functional View for Zero-Shot NAS
(
In-depth presentation
)
>
|
Ming Lin 🔗 |
Wed 1:55 p.m. - 2:05 p.m.
|
Break
|
🔗 |
Wed 2:05 p.m. - 2:45 p.m.
|
Understanding and Accelerating NAS with Theory-Grounded Metrics
(
In-depth presentation
)
>
|
Atlas Wang 🔗 |
Wed 2:45 p.m. - 3:00 p.m.
|
Break
|
🔗 |
Wed 3:00 p.m. - 3:40 p.m.
|
Network topology influence on gradient propagation and model performance
(
In-depth presentation
)
>
|
Kartikeya Bhardwaj 🔗 |
Wed 3:40 p.m. - 3:50 p.m.
|
Break
|
🔗 |
Wed 3:50 p.m. - 4:30 p.m.
|
Fast NAS with hardware co-optimization
(
In-depth presentation
)
>
|
Radu Marculescu 🔗 |
Wed 4:30 p.m. - 4:45 p.m.
|
Final thoughts
(
Summary
)
>
|
Radu Marculescu 🔗 |
Wed 4:45 p.m. - 5:00 p.m.
|
Q&A
(
Discussion
)
>
|
Ming Lin · Atlas Wang · Kartikeya Bhardwaj · Radu Marculescu 🔗 |