Not So Cooperative Caching(NSCC) considers a network comprised of selfish nodes; each is with caching capability and an objective of reducing its own access cost by fetching data from its local cache or from neighbori...Not So Cooperative Caching(NSCC) considers a network comprised of selfish nodes; each is with caching capability and an objective of reducing its own access cost by fetching data from its local cache or from neighboring caches. These nodes would cooperate in caching and share cached content if and only if they each benefit. The challenges are to determine what objects to cache at each node and to implement the system in the context of Information Centric Networking(ICN). This work includes both a solution for the NSCC problem and a design and implementation of an NSCC system in Named Data Networking(NDN), a large effort that exemplifies ICN. Our design applies NDN synchronization protocol to facilitate the information exchange among nodes, adopts group key encryption to control data access within the NSCC group, and offers an error checker to detect error events in the system. Our approach is validated by deploying the system we developed on Planet Lab.展开更多
With the widespread use of streaming media application on the Internet, a significant change in Internet workload will be provoked. Caching is one of the applied techniques for enhancing the scalability of streaming s...With the widespread use of streaming media application on the Internet, a significant change in Internet workload will be provoked. Caching is one of the applied techniques for enhancing the scalability of streaming system and reducing the workload of server/network. Aiming at the characteristics of broadband network in community, we propose a popularity-based server-proxy caching strategy for streaming medias, and implement the prototype of streaming proxy caching based on this strategy, using RTSP as control protocol and RTP for content transport. This system can play a role in decreasing server load, reducing the traffic from streaming server to proxy, and improving the start-up latency of the client. Key words streaming server - proxy - cache - streaming media - real time streaming protocol CLC number TP 302 - TP 333 Foundation item: Supported by the National High Technology Development 863 Program of China (2001AA111011).Biography: Tan Jin (1962-), male, Ph. D candidate, research direction: network communications, multimedia technologies, and web caching.展开更多
<div style="text-align:justify;"> <span style="font-family:Verdana;">There has been an explosion in the volume of data that is being accessed from the Internet. As a result, the risk of...<div style="text-align:justify;"> <span style="font-family:Verdana;">There has been an explosion in the volume of data that is being accessed from the Internet. As a result, the risk of a Web server being inundated with requests is ever-present. One approach to reducing the performance degradation that potentially comes from Web server overloading is to employ Web caching where data content is replicated in multiple locations. In this paper, we investigate the use of evolutionary algorithms to dynamically alter partition size in Web caches. We use established modeling techniques to compare the performance of our evolutionary algorithm to that found in statically-partitioned systems. Our results indicate that utilizing an evolutionary algorithm to dynamically alter partition sizes can lead to performance improvements especially in environments where the relative size of large to small pages is high.</span> </div>展开更多
In this paper,unmanned aerial vehicle(UAV)is adopted to serve as aerial base station(ABS)and mobile edge computing(MEC)platform for wire-less communication systems.When Internet of Things devices(IoTDs)cannot cope wit...In this paper,unmanned aerial vehicle(UAV)is adopted to serve as aerial base station(ABS)and mobile edge computing(MEC)platform for wire-less communication systems.When Internet of Things devices(IoTDs)cannot cope with computation-intensive and/or time-sensitive tasks,part of tasks is offloaded to the UAV side,and UAV process them with its own computing resources and caching resources.Thus,the burden of IoTDs gets relieved under the satisfaction of the quality of service(QoS)require-ments.However,owing to the limited resources of UAV,the cost of whole system,i.e.,that is defined as the weighted sum of energy consumption and time de-lay with caching,should be further optimized while the objective function and the constraints are non-convex.Therefore,we first jointly optimize commu-nication resources B,computing resources F and of-floading rates X with alternating iteration and convex optimization method,and then determine the value of caching decision Y with branch-and-bound(BB)al-gorithm.Numerical results show that UAV assisting partial task offloading with content caching is supe-rior to local computing and full offloading mechanism without caching,and meanwhile the cost of whole sys-tem gets further optimized with our proposed scheme.展开更多
The demands on conventional communication networks are increasing rapidly because of the exponential expansion of connected multimedia content.In light of the data-centric aspect of contemporary communication,the info...The demands on conventional communication networks are increasing rapidly because of the exponential expansion of connected multimedia content.In light of the data-centric aspect of contemporary communication,the information-centric network(ICN)paradigm offers hope for a solution by emphasizing content retrieval by name instead of location.If 5G networks are to meet the expected data demand surge from expanded connectivity and Internet of Things(IoT)devices,then effective caching solutions will be required tomaximize network throughput andminimize the use of resources.Hence,an ICN-based Cooperative Caching(ICN-CoC)technique has been used to select a cache by considering cache position,content attractiveness,and rate prediction.The findings show that utilizing our suggested approach improves caching regarding the Cache Hit Ratio(CHR)of 84.3%,Average Hop Minimization Ratio(AHMR)of 89.5%,and Mean Access Latency(MAL)of 0.4 s.Within a framework,it suggests improved caching strategies to handle the difficulty of effectively controlling data consumption in 5G networks.These improvements aim to make the network run more smoothly by enhancing content delivery,decreasing latency,and relieving congestion.By improving 5G communication systems’capacity tomanage the demands faced by modern data-centric applications,the research ultimately aids in advancement.展开更多
With the rapid development of 5G technology,the proportion of video traffic on the Internet is increasing,bringing pressure on the network infrastructure.Edge computing technology provides a feasible solution for opti...With the rapid development of 5G technology,the proportion of video traffic on the Internet is increasing,bringing pressure on the network infrastructure.Edge computing technology provides a feasible solution for optimizing video content distribution.However,the limited edge node cache capacity and dynamic user requests make edge caching more complex.Therefore,we propose a recommendation-driven edge Caching network architecture for the Full life cycle of video streaming(FlyCache)designed to improve users’Quality of Experience(QoE)and reduce backhaul traffic consumption.FlyCache implements intelligent caching management across three key stages:before-playback,during-playback,and after-playback.Specifically,we introduce a cache placement policy for the before-playback stage,a dynamic prefetching and cache admission policy for the during-playback stage,and a progressive cache eviction policy for the after-playback stage.To validate the effectiveness of FlyCache,we developed a user behavior-driven edge caching simulation framework incorporating recommendation mechanisms.Experiments conducted on the MovieLens and synthetic datasets demonstrate that FlyCache outperforms other caching strategies in terms of byte hit rate,backhaul traffic,and delayed startup rate.展开更多
Existing wireless networks are flooded with video data transmissions,and the demand for high-speed and low-latency video services continues to surge.This has brought with it challenges to networks in the form of conge...Existing wireless networks are flooded with video data transmissions,and the demand for high-speed and low-latency video services continues to surge.This has brought with it challenges to networks in the form of congestion as well as the need for more resources and more dedicated caching schemes.Recently,Multi-access Edge Computing(MEC)-enabled heterogeneous networks,which leverage edge caches for proximity delivery,have emerged as a promising solution to all of these problems.Designing an effective edge caching scheme is critical to its success,however,in the face of limited resources.We propose a novel Knowledge Graph(KG)-based Dueling Deep Q-Network(KG-DDQN)for cooperative caching in MEC-enabled heterogeneous networks.The KGDDQN scheme leverages a KG to uncover video relations,providing valuable insights into user preferences for the caching scheme.Specifically,the KG guides the selection of related videos as caching candidates(i.e.,actions in the DDQN),thus providing a rich reference for implementing a personalized caching scheme while also improving the decision efficiency of the DDQN.Extensive simulation results validate the convergence effectiveness of the KG-DDQN,and it also outperforms baselines regarding cache hit rate and service delay.展开更多
Efficient edge caching is essential for maximizing utility in video streaming systems,especially under constraints such as limited storage capacity and dynamically fluctuating content popularity.Utility,defined as the...Efficient edge caching is essential for maximizing utility in video streaming systems,especially under constraints such as limited storage capacity and dynamically fluctuating content popularity.Utility,defined as the benefit obtained per unit of cache bandwidth usage,degrades when static or greedy caching strategies fail to adapt to changing demand patterns.To address this,we propose a deep reinforcement learning(DRL)-based caching framework built upon the proximal policy optimization(PPO)algorithm.Our approach formulates edge caching as a sequential decision-making problem and introduces a reward model that balances cache hit performance and utility by prioritizing high-demand,high-quality content while penalizing degraded quality delivery.We construct a realistic synthetic dataset that captures both temporal variations and shifting content popularity to validate our model.Experimental results demonstrate that our proposed method improves utility by up to 135.9%and achieves an average improvement of 22.6%compared to traditional greedy algorithms and long short-term memory(LSTM)-based prediction models.Moreover,our method consistently performs well across a variety of utility functions,workload distributions,and storage limitations,underscoring its adaptability and robustness in dynamic video caching environments.展开更多
Vehicular networks enable seamless connectivity for exchanging emergency and infotainment content.However,retrieving infotainment data from remote servers often introduces high delays,degrading the Quality of Service(...Vehicular networks enable seamless connectivity for exchanging emergency and infotainment content.However,retrieving infotainment data from remote servers often introduces high delays,degrading the Quality of Service(QoS).To overcome this,caching frequently requested content at fog-enabled Road Side Units(RSUs)reduces communication latency.Yet,the limited caching capacity of RSUs makes it impractical to store all contents with varying sizes and popularity.This research proposes an efficient content caching algorithm that adapts to dynamic vehicular demands on highways to maximize request satisfaction.The scheme is evaluated against Intelligent Content Caching(ICC)and Random Caching(RC).The obtained results show that our proposed scheme entertains more contentrequesting vehicles as compared to ICC and RC,with 33%and 41%more downloaded data in 28%and 35%less amount of time from ICC and RC schemes,respectively.展开更多
Increasing reliance on large-scale AI models has led to rising demand for intelligent services.The centralized cloud computing approach has limitations in terms of data transfer efficiency and response time,and as a r...Increasing reliance on large-scale AI models has led to rising demand for intelligent services.The centralized cloud computing approach has limitations in terms of data transfer efficiency and response time,and as a result many service providers have begun to deploy edge servers to cache intelligent services in order to reduce transmission delay and communication energy consumption.However,finding the optimal service caching strategy remains a significant challenge due to the stochastic nature of service requests and the bulky nature of intelligent services.To deal with this,we propose a distributed service caching scheme integrating deep reinforcement learning(DRL)with mobility prediction,which we refer to as DSDM.Specifically,we employ the D3QN(Deep Double Dueling Q-Network)framework to integrate Long Short-Term Memory(LSTM)predicted mobile device locations into the service caching replacement algorithm and adopt the distributed multi-agent approach for learning and training.Experimental results demonstrate that DSDM achieves significant performance improvements in reducing communication energy consumption compared to traditional methods across various scenarios.展开更多
With the rapid development of generative artificial intelligence technology,the traditional cloud-based centralized model training and inference face significant limitations due to high transmission latency and costs,...With the rapid development of generative artificial intelligence technology,the traditional cloud-based centralized model training and inference face significant limitations due to high transmission latency and costs,which restrict user-side in-situ Artificial Intelligence Generated Content(AIGC)service requests.To this end,we propose the Edge Artificial Intelligence Generated Content(Edge AIGC)framework,which can effectively address the challenges of cloud computing by implementing in-situ processing of services close to the data source through edge computing.However,AIGC models usually have a large parameter scale and complex computing requirements,which poses a huge challenge to the storage and computing resources of edge devices.This paper focuses on the edge intelligence model caching and resource allocation problems in the Edge AIGC framework,aiming to improve the cache hit rate and resource utilization of edge devices for models by optimizing the model caching strategy and resource allocation scheme,and realize in-situ AIGC service processing.With the optimization objectives of minimizing service request response time and execution cost in resource-constrained environments,we employ the Twin Delayed Deep Deterministic Policy Gradient algorithm for optimization.Experimental results show that,compared with other methods,our model caching and resource allocation strategies can effectively improve the cache hit rate by at least 41.06%and reduce the response cost as well.展开更多
In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic natu...In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic nature of user locations and the potential of device-to-device(D2D)cooperative caching,limiting the reduction of transmission latency.To address this issue,this paper proposes a joint optimization scheme for edge caching that integrates user mobility prediction with deep reinforcement learning.First,a Transformer-based geolocation prediction model is designed,leveraging multi-head attention mechanisms to capture correlations in historical user trajectories for accurate future location prediction.Then,within a three-tier heterogeneous network,we formulate a latency minimization problem under a D2D cooperative caching architecture and develop a mobility-aware Deep Q-Network(DQN)caching strategy.This strategy takes predicted location information as state input and dynamically adjusts the content distribution across small base stations(SBSs)andmobile users(MUs)to reduce end-to-end delay inmulti-hop content retrieval.Simulation results show that the proposed DQN-based method outperforms other baseline strategies across variousmetrics,achieving a 17.2%reduction in transmission delay compared to DQNmethods withoutmobility integration,thus validating the effectiveness of the joint optimization of location prediction and caching decisions.展开更多
Named data networking(NDNs)is an idealized deployment of information-centric networking(ICN)that has attracted attention from scientists and scholars worldwide.A distributed in-network caching scheme can efficiently r...Named data networking(NDNs)is an idealized deployment of information-centric networking(ICN)that has attracted attention from scientists and scholars worldwide.A distributed in-network caching scheme can efficiently realize load balancing.However,such a ubiquitous caching approach may cause problems including duplicate caching and low data diversity,thus reducing the caching efficiency of NDN routers.To mitigate these caching problems and improve the NDN caching efficiency,in this paper,a hierarchical-based sequential caching(HSC)scheme is proposed.In this scheme,the NDN routers in the data transmission path are divided into various levels and data with different request frequencies are cached in distinct router levels.The aim is to cache data with high request frequencies in the router that is closest to the content requester to increase the response probability of the nearby data,improve the data caching efficiency of named data networks,shorten the response time,and reduce cache redundancy.Simulation results show that this scheme can effectively improve the cache hit rate(CHR)and reduce the average request delay(ARD)and average route hop(ARH).展开更多
Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrain...Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.展开更多
Recently, due to the deployment flexibility of unmanned aerial vehicles(UAVs), UAV-assisted mobile relay communication system has been widely used in the maritime communication. However, the performance of UAV-assiste...Recently, due to the deployment flexibility of unmanned aerial vehicles(UAVs), UAV-assisted mobile relay communication system has been widely used in the maritime communication. However, the performance of UAV-assisted mobile relay communication system is limited by the capacity of wireless backhaul link between base station and UAV. In this paper, we consider a caching UAV-assisted decode-and-forward relay communication system in a downlink maritime communication. For the general case with multiple users, the optimal placement of UAV is obtained by solving the average achievable rate maximization problem through the one-dimensional linear search. For a special case with single user, we derive a semi closedform expression of the optimal placement of UAV. Simulation results confirm the accuracy of analytical results and show that the optimal placement of UAV and the average achievable rate significantly depend on the cache capacity at UAV. We also show the difference between the performances of the air-to-ground model and the air-to-sea model.展开更多
基金sponsored by the National Grand Fundamental Research 973 program of China under Grant No.2009CB320505the National Nature Science Foundation of China under Grant No. 60973123+1 种基金the Technology Support Program (Industry) of Jiangsu under Grant No.BE2011173Prospective Research Project on Future Networks of Jiangsu Future Networks Innovation Institute under Grant No.BY2013095-5-03
文摘Not So Cooperative Caching(NSCC) considers a network comprised of selfish nodes; each is with caching capability and an objective of reducing its own access cost by fetching data from its local cache or from neighboring caches. These nodes would cooperate in caching and share cached content if and only if they each benefit. The challenges are to determine what objects to cache at each node and to implement the system in the context of Information Centric Networking(ICN). This work includes both a solution for the NSCC problem and a design and implementation of an NSCC system in Named Data Networking(NDN), a large effort that exemplifies ICN. Our design applies NDN synchronization protocol to facilitate the information exchange among nodes, adopts group key encryption to control data access within the NSCC group, and offers an error checker to detect error events in the system. Our approach is validated by deploying the system we developed on Planet Lab.
文摘With the widespread use of streaming media application on the Internet, a significant change in Internet workload will be provoked. Caching is one of the applied techniques for enhancing the scalability of streaming system and reducing the workload of server/network. Aiming at the characteristics of broadband network in community, we propose a popularity-based server-proxy caching strategy for streaming medias, and implement the prototype of streaming proxy caching based on this strategy, using RTSP as control protocol and RTP for content transport. This system can play a role in decreasing server load, reducing the traffic from streaming server to proxy, and improving the start-up latency of the client. Key words streaming server - proxy - cache - streaming media - real time streaming protocol CLC number TP 302 - TP 333 Foundation item: Supported by the National High Technology Development 863 Program of China (2001AA111011).Biography: Tan Jin (1962-), male, Ph. D candidate, research direction: network communications, multimedia technologies, and web caching.
文摘<div style="text-align:justify;"> <span style="font-family:Verdana;">There has been an explosion in the volume of data that is being accessed from the Internet. As a result, the risk of a Web server being inundated with requests is ever-present. One approach to reducing the performance degradation that potentially comes from Web server overloading is to employ Web caching where data content is replicated in multiple locations. In this paper, we investigate the use of evolutionary algorithms to dynamically alter partition size in Web caches. We use established modeling techniques to compare the performance of our evolutionary algorithm to that found in statically-partitioned systems. Our results indicate that utilizing an evolutionary algorithm to dynamically alter partition sizes can lead to performance improvements especially in environments where the relative size of large to small pages is high.</span> </div>
基金supported by National Natural Science Foundation of China(No.61821001)Science and Technology Key Project of Guangdong Province,China(2019B010157001).
文摘In this paper,unmanned aerial vehicle(UAV)is adopted to serve as aerial base station(ABS)and mobile edge computing(MEC)platform for wire-less communication systems.When Internet of Things devices(IoTDs)cannot cope with computation-intensive and/or time-sensitive tasks,part of tasks is offloaded to the UAV side,and UAV process them with its own computing resources and caching resources.Thus,the burden of IoTDs gets relieved under the satisfaction of the quality of service(QoS)require-ments.However,owing to the limited resources of UAV,the cost of whole system,i.e.,that is defined as the weighted sum of energy consumption and time de-lay with caching,should be further optimized while the objective function and the constraints are non-convex.Therefore,we first jointly optimize commu-nication resources B,computing resources F and of-floading rates X with alternating iteration and convex optimization method,and then determine the value of caching decision Y with branch-and-bound(BB)al-gorithm.Numerical results show that UAV assisting partial task offloading with content caching is supe-rior to local computing and full offloading mechanism without caching,and meanwhile the cost of whole sys-tem gets further optimized with our proposed scheme.
基金New Brunswick Innovation Foundation(NBIF)for the financial support of the global project.
文摘The demands on conventional communication networks are increasing rapidly because of the exponential expansion of connected multimedia content.In light of the data-centric aspect of contemporary communication,the information-centric network(ICN)paradigm offers hope for a solution by emphasizing content retrieval by name instead of location.If 5G networks are to meet the expected data demand surge from expanded connectivity and Internet of Things(IoT)devices,then effective caching solutions will be required tomaximize network throughput andminimize the use of resources.Hence,an ICN-based Cooperative Caching(ICN-CoC)technique has been used to select a cache by considering cache position,content attractiveness,and rate prediction.The findings show that utilizing our suggested approach improves caching regarding the Cache Hit Ratio(CHR)of 84.3%,Average Hop Minimization Ratio(AHMR)of 89.5%,and Mean Access Latency(MAL)of 0.4 s.Within a framework,it suggests improved caching strategies to handle the difficulty of effectively controlling data consumption in 5G networks.These improvements aim to make the network run more smoothly by enhancing content delivery,decreasing latency,and relieving congestion.By improving 5G communication systems’capacity tomanage the demands faced by modern data-centric applications,the research ultimately aids in advancement.
基金supported by the National Natural Science Foundation of China(NSFC)[Grant No.62072469].
文摘With the rapid development of 5G technology,the proportion of video traffic on the Internet is increasing,bringing pressure on the network infrastructure.Edge computing technology provides a feasible solution for optimizing video content distribution.However,the limited edge node cache capacity and dynamic user requests make edge caching more complex.Therefore,we propose a recommendation-driven edge Caching network architecture for the Full life cycle of video streaming(FlyCache)designed to improve users’Quality of Experience(QoE)and reduce backhaul traffic consumption.FlyCache implements intelligent caching management across three key stages:before-playback,during-playback,and after-playback.Specifically,we introduce a cache placement policy for the before-playback stage,a dynamic prefetching and cache admission policy for the during-playback stage,and a progressive cache eviction policy for the after-playback stage.To validate the effectiveness of FlyCache,we developed a user behavior-driven edge caching simulation framework incorporating recommendation mechanisms.Experiments conducted on the MovieLens and synthetic datasets demonstrate that FlyCache outperforms other caching strategies in terms of byte hit rate,backhaul traffic,and delayed startup rate.
基金supported by the National Natural Science Foundation of China(Nos.62201419,62372357)the Natural Science Foundation of Chongqing(CSTB2023NSCQ-LMX0032)the ISN State Key Laboratory.
文摘Existing wireless networks are flooded with video data transmissions,and the demand for high-speed and low-latency video services continues to surge.This has brought with it challenges to networks in the form of congestion as well as the need for more resources and more dedicated caching schemes.Recently,Multi-access Edge Computing(MEC)-enabled heterogeneous networks,which leverage edge caches for proximity delivery,have emerged as a promising solution to all of these problems.Designing an effective edge caching scheme is critical to its success,however,in the face of limited resources.We propose a novel Knowledge Graph(KG)-based Dueling Deep Q-Network(KG-DDQN)for cooperative caching in MEC-enabled heterogeneous networks.The KGDDQN scheme leverages a KG to uncover video relations,providing valuable insights into user preferences for the caching scheme.Specifically,the KG guides the selection of related videos as caching candidates(i.e.,actions in the DDQN),thus providing a rich reference for implementing a personalized caching scheme while also improving the decision efficiency of the DDQN.Extensive simulation results validate the convergence effectiveness of the KG-DDQN,and it also outperforms baselines regarding cache hit rate and service delay.
文摘Efficient edge caching is essential for maximizing utility in video streaming systems,especially under constraints such as limited storage capacity and dynamically fluctuating content popularity.Utility,defined as the benefit obtained per unit of cache bandwidth usage,degrades when static or greedy caching strategies fail to adapt to changing demand patterns.To address this,we propose a deep reinforcement learning(DRL)-based caching framework built upon the proximal policy optimization(PPO)algorithm.Our approach formulates edge caching as a sequential decision-making problem and introduces a reward model that balances cache hit performance and utility by prioritizing high-demand,high-quality content while penalizing degraded quality delivery.We construct a realistic synthetic dataset that captures both temporal variations and shifting content popularity to validate our model.Experimental results demonstrate that our proposed method improves utility by up to 135.9%and achieves an average improvement of 22.6%compared to traditional greedy algorithms and long short-term memory(LSTM)-based prediction models.Moreover,our method consistently performs well across a variety of utility functions,workload distributions,and storage limitations,underscoring its adaptability and robustness in dynamic video caching environments.
基金supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(grant number IMSIU-DDRSP2504).
文摘Vehicular networks enable seamless connectivity for exchanging emergency and infotainment content.However,retrieving infotainment data from remote servers often introduces high delays,degrading the Quality of Service(QoS).To overcome this,caching frequently requested content at fog-enabled Road Side Units(RSUs)reduces communication latency.Yet,the limited caching capacity of RSUs makes it impractical to store all contents with varying sizes and popularity.This research proposes an efficient content caching algorithm that adapts to dynamic vehicular demands on highways to maximize request satisfaction.The scheme is evaluated against Intelligent Content Caching(ICC)and Random Caching(RC).The obtained results show that our proposed scheme entertains more contentrequesting vehicles as compared to ICC and RC,with 33%and 41%more downloaded data in 28%and 35%less amount of time from ICC and RC schemes,respectively.
基金supported by the National Natural Science Foundation of China under grants No.92267104 and 62372242。
文摘Increasing reliance on large-scale AI models has led to rising demand for intelligent services.The centralized cloud computing approach has limitations in terms of data transfer efficiency and response time,and as a result many service providers have begun to deploy edge servers to cache intelligent services in order to reduce transmission delay and communication energy consumption.However,finding the optimal service caching strategy remains a significant challenge due to the stochastic nature of service requests and the bulky nature of intelligent services.To deal with this,we propose a distributed service caching scheme integrating deep reinforcement learning(DRL)with mobility prediction,which we refer to as DSDM.Specifically,we employ the D3QN(Deep Double Dueling Q-Network)framework to integrate Long Short-Term Memory(LSTM)predicted mobile device locations into the service caching replacement algorithm and adopt the distributed multi-agent approach for learning and training.Experimental results demonstrate that DSDM achieves significant performance improvements in reducing communication energy consumption compared to traditional methods across various scenarios.
基金supported in part by the Shandong Provincial Natural Science Foundation under Grants ZR2023LZH017,ZR2022LZH015 and ZR2024MF066the National Natural Science Foundation of China under Grant 62471493+1 种基金the Tertiary Education Scientific research project of Guangzhou Municipal Education Bureau under Grant 2024312246the Guangzhou Higher Education Teaching Quality and Teaching Reform Project under Grant 2023KCJJD002。
文摘With the rapid development of generative artificial intelligence technology,the traditional cloud-based centralized model training and inference face significant limitations due to high transmission latency and costs,which restrict user-side in-situ Artificial Intelligence Generated Content(AIGC)service requests.To this end,we propose the Edge Artificial Intelligence Generated Content(Edge AIGC)framework,which can effectively address the challenges of cloud computing by implementing in-situ processing of services close to the data source through edge computing.However,AIGC models usually have a large parameter scale and complex computing requirements,which poses a huge challenge to the storage and computing resources of edge devices.This paper focuses on the edge intelligence model caching and resource allocation problems in the Edge AIGC framework,aiming to improve the cache hit rate and resource utilization of edge devices for models by optimizing the model caching strategy and resource allocation scheme,and realize in-situ AIGC service processing.With the optimization objectives of minimizing service request response time and execution cost in resource-constrained environments,we employ the Twin Delayed Deep Deterministic Policy Gradient algorithm for optimization.Experimental results show that,compared with other methods,our model caching and resource allocation strategies can effectively improve the cache hit rate by at least 41.06%and reduce the response cost as well.
基金supported by the Liaoning Provincial Education Department Fund,grant number JYTZD2023083.
文摘In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic nature of user locations and the potential of device-to-device(D2D)cooperative caching,limiting the reduction of transmission latency.To address this issue,this paper proposes a joint optimization scheme for edge caching that integrates user mobility prediction with deep reinforcement learning.First,a Transformer-based geolocation prediction model is designed,leveraging multi-head attention mechanisms to capture correlations in historical user trajectories for accurate future location prediction.Then,within a three-tier heterogeneous network,we formulate a latency minimization problem under a D2D cooperative caching architecture and develop a mobility-aware Deep Q-Network(DQN)caching strategy.This strategy takes predicted location information as state input and dynamically adjusts the content distribution across small base stations(SBSs)andmobile users(MUs)to reduce end-to-end delay inmulti-hop content retrieval.Simulation results show that the proposed DQN-based method outperforms other baseline strategies across variousmetrics,achieving a 17.2%reduction in transmission delay compared to DQNmethods withoutmobility integration,thus validating the effectiveness of the joint optimization of location prediction and caching decisions.
基金supported in part by the National Natural Science Foundation of China under Grant 61972424 and 62372479in part by the High Value Intellectual Property Cultivation Project of Hubei Province,China,under grant D2021002094+1 种基金in part by JSPS KAKENHI under Grants JP16K00117 and JP19K20250in part by the Leading Initiative for Excellent Young Researchers(LEADER),MEXT,Japan,and KDDI Foundation.
文摘Named data networking(NDNs)is an idealized deployment of information-centric networking(ICN)that has attracted attention from scientists and scholars worldwide.A distributed in-network caching scheme can efficiently realize load balancing.However,such a ubiquitous caching approach may cause problems including duplicate caching and low data diversity,thus reducing the caching efficiency of NDN routers.To mitigate these caching problems and improve the NDN caching efficiency,in this paper,a hierarchical-based sequential caching(HSC)scheme is proposed.In this scheme,the NDN routers in the data transmission path are divided into various levels and data with different request frequencies are cached in distinct router levels.The aim is to cache data with high request frequencies in the router that is closest to the content requester to increase the response probability of the nearby data,improve the data caching efficiency of named data networks,shorten the response time,and reduce cache redundancy.Simulation results show that this scheme can effectively improve the cache hit rate(CHR)and reduce the average request delay(ARD)and average route hop(ARH).
基金supported by Key Science and Technology Program of Henan Province,China(Grant Nos.242102210147,242102210027)Fujian Province Young and Middle aged Teacher Education Research Project(Science and Technology Category)(No.JZ240101)(Corresponding author:Dong Yuan).
文摘Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.
基金supported in part by the Natural Science Foundation of China under Grant U1805262,61671251,61871446,61701118,61871131,and 61404130218the Natural Science Foundation of Fujian Province under Grant 2018J05101。
文摘Recently, due to the deployment flexibility of unmanned aerial vehicles(UAVs), UAV-assisted mobile relay communication system has been widely used in the maritime communication. However, the performance of UAV-assisted mobile relay communication system is limited by the capacity of wireless backhaul link between base station and UAV. In this paper, we consider a caching UAV-assisted decode-and-forward relay communication system in a downlink maritime communication. For the general case with multiple users, the optimal placement of UAV is obtained by solving the average achievable rate maximization problem through the one-dimensional linear search. For a special case with single user, we derive a semi closedform expression of the optimal placement of UAV. Simulation results confirm the accuracy of analytical results and show that the optimal placement of UAV and the average achievable rate significantly depend on the cache capacity at UAV. We also show the difference between the performances of the air-to-ground model and the air-to-sea model.