TY - GEN
T1 - Distributionally robust edge learning with dirichlet process prior
AU - Zhang, Zhaofeng
AU - Chen, Yue
AU - Zhang, Junshan
N1 - Funding Information:
This work is supported in part by NSF under Grant CPS-1739344, ARO under grant W911NF-16-1-0448, and the DTRA under Grant HDTRA1-13-1-0029.
Publisher Copyright:
© 2020 IEEE
PY - 2020/11
Y1 - 2020/11
N2 - In order to meet the real-time performance requirements, intelligent decisions in many IoT applications must take place right here right now at the network edge. The conventional cloud-based learning approach would not be able to keep up with the demands in achieving edge intelligence in these applications. Nevertheless, pushing the artificial intelligence (AI) frontier to achieve edge intelligence is highly nontrivial due to the constrained computing resources and limited training data at the network edge. To tackle these challenges, we develop a distributionally robust optimization (DRO)-based edge learning algorithm, where the uncertainty model is constructed to foster the synergy of cloud knowledge transfer and local training. Specifically, the knowledge transferred from the cloud is in the form of a Dirichlet process prior distribution for the edge model parameters, and the edge device further constructs an uncertainty set centered around the empirical distribution of its local samples to capture the information of local data processing. The edge learning DRO problem, subject to the above two distributional uncertainty constraints, is then recast as an equivalent single-layer optimization problem using a duality approach. We then use an Expectation-Maximization (EM) algorithm-inspired method to derive a convex relaxation, based on which we devise algorithms to learn the edge model parameters. Finally, extensive experiments are implemented to showcase the performance gain over standard learning approaches using local edge data only.
AB - In order to meet the real-time performance requirements, intelligent decisions in many IoT applications must take place right here right now at the network edge. The conventional cloud-based learning approach would not be able to keep up with the demands in achieving edge intelligence in these applications. Nevertheless, pushing the artificial intelligence (AI) frontier to achieve edge intelligence is highly nontrivial due to the constrained computing resources and limited training data at the network edge. To tackle these challenges, we develop a distributionally robust optimization (DRO)-based edge learning algorithm, where the uncertainty model is constructed to foster the synergy of cloud knowledge transfer and local training. Specifically, the knowledge transferred from the cloud is in the form of a Dirichlet process prior distribution for the edge model parameters, and the edge device further constructs an uncertainty set centered around the empirical distribution of its local samples to capture the information of local data processing. The edge learning DRO problem, subject to the above two distributional uncertainty constraints, is then recast as an equivalent single-layer optimization problem using a duality approach. We then use an Expectation-Maximization (EM) algorithm-inspired method to derive a convex relaxation, based on which we devise algorithms to learn the edge model parameters. Finally, extensive experiments are implemented to showcase the performance gain over standard learning approaches using local edge data only.
KW - Dirichlet process
KW - Distributionally robust optimization
KW - Edge learning
KW - Wasserstein distance
UR - http://www.scopus.com/inward/record.url?scp=85101986492&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85101986492&partnerID=8YFLogxK
U2 - 10.1109/ICDCS47774.2020.00016
DO - 10.1109/ICDCS47774.2020.00016
M3 - Conference contribution
AN - SCOPUS:85101986492
T3 - Proceedings - International Conference on Distributed Computing Systems
SP - 798
EP - 808
BT - Proceedings - 2020 IEEE 40th International Conference on Distributed Computing Systems, ICDCS 2020
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 40th IEEE International Conference on Distributed Computing Systems, ICDCS 2020
Y2 - 29 November 2020 through 1 December 2020
ER -