This workshop seeks to bring ML and wireless networking experts together to identify interdisciplinary approaches to evolve ML algorithms for and over communication networks that operate under constrained resources, including time, labeling, and computational capacity constraints. The workshop will provide a unique opportunity to expose the MLSys community to the challenges and opportunities of integrating ML methods into resource-constrained communication networks. It will also highlight emerging trends in ML with limited resources and their implications for the design and operation of next-generation communication networks.
We are seeking original submissions in topics including, but not limited to:
- Learning in wireless networks with limited training data
- Multi-agent federated/distributed learning with low computational and communication resources
- Communicated data compression for network-wide task completion
- Online learning with wireless latency constraints
- Learning in wireless networks with privacy constraints
- Few-shot learning and adaptation in wireless environments
- Datasets and benchmarks for resource-constrained learning in wireless networks
Thu 5:50 a.m. - 6:00 a.m.
|
Opening Remarks (Ervin Moore)
(
Introduction
)
|
Ervin Moore 🔗 |
Thu 6:00 a.m. - 6:30 a.m.
|
Invited Talk (Hyeji Kim)
(
Invited Talk
)
|
Hyeji Kim 🔗 |
Thu 6:30 a.m. - 7:00 a.m.
|
Invited Talk (Osvaldo Simeone)
(
Invited Talk
)
|
Osvaldo Simeone 🔗 |
Thu 7:00 a.m. - 7:30 a.m.
|
Invited Talk (Alejandro Ribeiro)
(
Invited Talk
)
|
Alejandro Ribeiro 🔗 |
Thu 7:30 a.m. - 7:45 a.m.
|
Break
|
🔗 |
Thu 7:45 a.m. - 8:15 a.m.
|
Invited Talk (Tara Javidi)
(
Invited Talk
)
|
Tara Javidi 🔗 |
Thu 8:15 a.m. - 8:45 a.m.
|
Invited Talk (Salman Avestimehr)
(
Invited Talk
)
|
Salman Avestimehr 🔗 |
Thu 8:45 a.m. - 9:45 a.m.
|
Lunch Break
|
🔗 |
Thu 9:45 a.m. - 10:00 a.m.
|
DHA-FL: Enabling Efficient and Effective AIoT via Decentralized Hierarchical Asynchronous Federated Learning
(
Oral
)
link »
The challenges of scalability, robustness, and resilience to slow devices have posed significant obstacles to the effective and efficient implementation of Federated Learning (FL), a crucial technology for the emerging Artificial Intelligence of Things (AIoT). This paper proposes a solution to these challenges with the introduction of a Decentralized Hierarchical Asynchronous Federated Learning Scheme (DHA-FL). This scheme utilizes a hierarchical edge computing architecture, enabling a two-stage model aggregation paradigm that significantly enhances system scalability. To further enhance system robustness, decentralized asynchronous model aggregation is adopted among edge servers to prevent single node failures while mitigating the impact of slow devices or stragglers. Our experiments, conducted on a live wireless multi-hop IoT testbed, demonstrate that DHA-FL can achieve convergence in approximately half the time compared to the centralized hierarchical approach. Moreover, it enables an even more significant convergence speed-up (up to 8x) over the classic FedAvg baseline when dealing with stragglers. |
Wesley Huff · pinyarash pinyoanuntapong · Ravikumar Balakrishnan · Hao Feng · Minwoo Lee · Pu Wang · Chen Chen 🔗 |
Thu 10:00 a.m. - 10:15 a.m.
|
Scalable Feature Compression for Edge-Assisted Object Detection Over Time-Varying Networks
(
Oral
)
link »
Split-computing has recently emerged as a paradigm for offloading computation of visual analytics models from low-powered mobile devices to edge or cloud servers, by which the mobiles execute part of the model and compress and send the intermediate features, and the servers complete the remaining model computation. Prior feature compression approaches train different compression models and possibly visual analytics models to reach different target bit rates. We propose a scalable compression model that compresses the intermediate features of the YOLO object detection model into a layered bitstream, which can be easily adapted to meet the rate constraint of a dynamic network. Our approach achieves comparable rate-accuracy performance compared to prior non-scalable compression approaches over a large bitrate range. |
Zhongzheng Yuan · Siddharth Garg · Elza Erkip · Yao Wang 🔗 |
Thu 10:15 a.m. - 10:30 a.m.
|
Break
|
🔗 |
Thu 10:30 a.m. - 11:00 a.m.
|
Invited Talk (Olga Galinina)
(
Invited Talk
)
|
Olga Galinina 🔗 |
Thu 11:00 a.m. - 11:15 a.m.
|
Break
|
🔗 |
Thu 11:15 a.m. - 11:30 a.m.
|
Trained-MPC: A Private Inference by Training-Based Multiparty Computation
(
Oral
)
link »
How can we perform inference on data using cloud servers without leaking any information to them? The answer lies in Trained-MPC, an innovative approach to inference privacy that can be applied to deep learning models. It relies on a cluster of servers, each running a learning model, which are fed with the client data added with strong noise. The noise is independent of user data, but dependent across the servers. The variance of the noise is set to be large enough to make the information leakage to the servers negligible. The dependency among the noise of the queries allows the parameters of the models running on different servers to be trained such that the client can mitigate the contribution of the noises by combining the outputs of the servers, and recover the final result with high accuracy and with a minor computational effort. In other words, in the proposed method, we develop a multiparty computation (MPC) by training for a specific inference task while avoiding the extensive communication overhead that MPC entails. Simulation results demonstrate Trained-MPC resolves the tension between privacy and accuracy while avoiding the computational and communication load needed in cryptography schemes. |
Hamidreza Ehteram · Mohammad Ali Maddah-Ali · Mahtab Mirmohseni 🔗 |
Thu 11:30 a.m. - 11:45 a.m.
|
Efficient and Light-Weight Federated Learning via Asynchronous Distributed Dropout
(
Oral
)
link »
Asynchronous learning protocols have regained attention lately, especially in the Federated Learning (FL) setup, where slower clients can severely impede the learning process. Herein, we propose \texttt{AsyncDrop}, a novel asynchronous FL framework that utilizes dropout regularization to handle device heterogeneity in distributed settings. Overall, \texttt{AsyncDrop} achieves better performance compared to state of the art asynchronous methodologies, while resulting in less communication and training time overheads. We implement our approach and compare it against other asynchronous baselines, both by design and by adapting existing synchronous FL algorithms to asynchronous scenarios. Empirically, \texttt{AsyncDrop} reduces the communication cost and training time, while matching or improving the final test accuracy in diverse non-i.i.d. FL scenarios. |
Chen Dun · Mirian Hipolito Garcia · Christopher Jermaine · Dimitrios Dimitriadis · Tasos Kyrillidis · Anastasios Kyrillidis 🔗 |
Thu 11:45 a.m. - 12:00 p.m.
|
Over-the-Air Federated TD Learning
(
Oral
)
link »
In recent years, federated learning has been widely studied to speed up various \textit{supervised} learning tasks at the wireless network edge under communication constraints. However, there is a lack of theoretical understanding as to whether similar speedups in sample complexity can be achieved for cooperative reinforcement learning (RL) problems subject to realistic communication models. To that end, we study a federated policy evaluation problem over wireless fading channels where, to update model parameters, a central server aggregates local temporal difference (TD) update directions from $N$ agents via analog over-the-air computation (OAC). We refer to this scheme as \texttt{OAC-FedTD} and provide a rigorous finite-time convergence analysis of its performance that accounts for linear function approximation, Markovian sampling, and channel-induced distortions and noise. Our analysis reveals the impact of the noisy fading channels on the convergence rate and establishes a linear convergence speedup w.r.t. the number of agents. As far as we are aware, this is the first non-asymptotic analysis of a cooperative RL setting under channel effects. Moreover, our proof leads to tighter bounds on the mixing time relative to existing work in federated RL (without channel effects); as such, it can be of independent interest.
|
Nicolò Dal Fabbro · Aritra Mitra · Robert Heath · Luca Schenato · George Pappas 🔗 |
Thu 12:00 p.m. - 12:05 p.m.
|
Closing Remarks (Ervin Moore)
(
Conclusion
)
|
Ervin Moore 🔗 |