In 6G,artificial intelligence represented by deep nerual network(DNN)will unleash its potential and empower IoT applications to transform into intelligent IoT applications.However,whole DNNbased inference is difficult...In 6G,artificial intelligence represented by deep nerual network(DNN)will unleash its potential and empower IoT applications to transform into intelligent IoT applications.However,whole DNNbased inference is difficult to carry out on resourceconstrained intelligent IoT devices and will suffer privacy leakage when offloading to the cloud or mobile edge computation server(MECs).In this paper,we formulate a privacy and delay dual-driven device-edge collaborative inference(P4DE-CI)system to preserve the privacy of raw data while accelerating the intelligent inference process,where the intelligent IoT devices run the front-end part of DNN model and the MECs execute the back-end part of DNN model.Considering three typical privacy leakage models and the end-to-end delay of collaborative DNN-based inference,we define a novel intelligent inference Quality of service(I2-QoS)metric as the weighted summation of the inference latency and privacy preservation level.Moreover,we propose a DDPG-based joint DNN model optimization and resource allocation algorithm to maximize I2-QoS,by optimizing the association relationship between intelligent IoT devices and MECs,the DNN model placement decision,and the DNN model partition decision.Experiments carried out on the AlexNet model reveal that the proposed algorithm has better performance in both privacy-preserving and inference-acceleration.展开更多
The lack of facial features caused by wearing masks degrades the performance of facial recognition systems.Traditional occluded face recognition methods cannot integrate the computational resources of the edge layer a...The lack of facial features caused by wearing masks degrades the performance of facial recognition systems.Traditional occluded face recognition methods cannot integrate the computational resources of the edge layer and the device layer.Besides,previous research fails to consider the facial characteristics including occluded and unoccluded parts.To solve the above problems,we put forward a device-edge collaborative occluded face recognition method based on cross-domain feature fusion.Specifically,the device-edge collaborative face recognition architecture gets the utmost out of maximizes device and edge resources for real-time occluded face recognition.Then,a cross-domain facial feature fusion method is presented which combines both the explicit domain and the implicit domain facial.Furthermore,a delay-optimized edge recognition task scheduling method is developed that comprehensively considers the task load,computational power,bandwidth,and delay tolerance constraints of the edge.This method can dynamically schedule face recognition tasks and minimize recognition delay while ensuring recognition accuracy.The experimental results show that the proposed method achieves an average gain of about 21%in recognition latency,while the accuracy of the face recognition task is basically the same compared to the baseline method.展开更多
By leveraging the data sample diversity,the early-exit network recently emerges as a prominent neural network architecture to accelerate the deep learning inference process.However,intermediate classifiers of the earl...By leveraging the data sample diversity,the early-exit network recently emerges as a prominent neural network architecture to accelerate the deep learning inference process.However,intermediate classifiers of the early exits introduce additional computation overhead,which is unfavorable for resource-constrained edge artificial intelligence(AI).In this paper,we propose an early exit prediction mechanism to reduce the on-device computation overhead in a device-edge co-inference system supported by early-exit networks.Specifically,we design a low-complexity module,namely the exit predictor,to guide some distinctly“hard”samples to bypass the computation of the early exits.Besides,considering the varying communication bandwidth,we extend the early exit prediction mechanism for latency-aware edge inference,which adapts the prediction thresholds of the exit predictor and the confidence thresholds of the early-exit network via a few simple regression models.Extensive experiment results demonstrate the effectiveness of the exit predictor in achieving a better tradeoff between accuracy and on-device computation overhead for early-exit networks.Besides,compared with the baseline methods,the proposed method for latency-aware edge inference attains higher inference accuracy under different bandwidth conditions.展开更多
基金supported by the National Natural Science Foundation of China(No.62201079)the Beijing Natural Science Foundation(No.L232051).
文摘In 6G,artificial intelligence represented by deep nerual network(DNN)will unleash its potential and empower IoT applications to transform into intelligent IoT applications.However,whole DNNbased inference is difficult to carry out on resourceconstrained intelligent IoT devices and will suffer privacy leakage when offloading to the cloud or mobile edge computation server(MECs).In this paper,we formulate a privacy and delay dual-driven device-edge collaborative inference(P4DE-CI)system to preserve the privacy of raw data while accelerating the intelligent inference process,where the intelligent IoT devices run the front-end part of DNN model and the MECs execute the back-end part of DNN model.Considering three typical privacy leakage models and the end-to-end delay of collaborative DNN-based inference,we define a novel intelligent inference Quality of service(I2-QoS)metric as the weighted summation of the inference latency and privacy preservation level.Moreover,we propose a DDPG-based joint DNN model optimization and resource allocation algorithm to maximize I2-QoS,by optimizing the association relationship between intelligent IoT devices and MECs,the DNN model placement decision,and the DNN model partition decision.Experiments carried out on the AlexNet model reveal that the proposed algorithm has better performance in both privacy-preserving and inference-acceleration.
基金supported by National Natural Science Foundation of China(61901071,61871062,61771082,U20A20157)Science and Natural Science Foundation of Chongqing,China(cstc2020jcyjzdxmX0024)+6 种基金University Innovation Research Group of Chongqing(CXQT20017)Program for Innovation Team Building at Institutions of Higher Education in Chongqing(CXTDX201601020)Natural Science Foundation of Chongqing,China(CSTB2022NSCQ-MSX0600)Youth Innovation Group Support Program of ICE Discipline of CQUPT(SCIE-QN-2022-04)Chongqing Municipal Technology Innovation and Application Development Special Key Project(cstc2020jscxdxwtBX0053)China Postdoctoral Science Foundation Project,China(2022MD723723)Chongqing Postdoctoral Research Project Special Funding,China(2023CQBSHTB3092)。
文摘The lack of facial features caused by wearing masks degrades the performance of facial recognition systems.Traditional occluded face recognition methods cannot integrate the computational resources of the edge layer and the device layer.Besides,previous research fails to consider the facial characteristics including occluded and unoccluded parts.To solve the above problems,we put forward a device-edge collaborative occluded face recognition method based on cross-domain feature fusion.Specifically,the device-edge collaborative face recognition architecture gets the utmost out of maximizes device and edge resources for real-time occluded face recognition.Then,a cross-domain facial feature fusion method is presented which combines both the explicit domain and the implicit domain facial.Furthermore,a delay-optimized edge recognition task scheduling method is developed that comprehensively considers the task load,computational power,bandwidth,and delay tolerance constraints of the edge.This method can dynamically schedule face recognition tasks and minimize recognition delay while ensuring recognition accuracy.The experimental results show that the proposed method achieves an average gain of about 21%in recognition latency,while the accuracy of the face recognition task is basically the same compared to the baseline method.
基金fund of the Hong Kong Polytechnic University(P0038174)。
文摘By leveraging the data sample diversity,the early-exit network recently emerges as a prominent neural network architecture to accelerate the deep learning inference process.However,intermediate classifiers of the early exits introduce additional computation overhead,which is unfavorable for resource-constrained edge artificial intelligence(AI).In this paper,we propose an early exit prediction mechanism to reduce the on-device computation overhead in a device-edge co-inference system supported by early-exit networks.Specifically,we design a low-complexity module,namely the exit predictor,to guide some distinctly“hard”samples to bypass the computation of the early exits.Besides,considering the varying communication bandwidth,we extend the early exit prediction mechanism for latency-aware edge inference,which adapts the prediction thresholds of the exit predictor and the confidence thresholds of the early-exit network via a few simple regression models.Extensive experiment results demonstrate the effectiveness of the exit predictor in achieving a better tradeoff between accuracy and on-device computation overhead for early-exit networks.Besides,compared with the baseline methods,the proposed method for latency-aware edge inference attains higher inference accuracy under different bandwidth conditions.