期刊文献+
共找到5,255篇文章
< 1 2 250 >
每页显示 20 50 100
R-DSP二级Cache的设计优化
1
作者 袁杰 吴丽娟 +1 位作者 杨德强 杨广林 《计算机工程与设计》 北大核心 2026年第2期568-575,共8页
通过比对R数字信号处理器(R-DSP)与TI-AWR2944对大型算法的处理周期数,发现R-DSP的二级Cache与TI-AWR2944的二级Cache存在着较大的性能差距。针对上述问题,提出采用门控脉冲时钟电路替代传统的跨时钟域方法,减少单次读写访问命中二级Ca... 通过比对R数字信号处理器(R-DSP)与TI-AWR2944对大型算法的处理周期数,发现R-DSP的二级Cache与TI-AWR2944的二级Cache存在着较大的性能差距。针对上述问题,提出采用门控脉冲时钟电路替代传统的跨时钟域方法,减少单次读写访问命中二级Cache所消耗的时钟周期数。此外,也在二级Cache中加入流水线结构,提高猝发访问二级Cache的效率。经UVM(universal verification methodology)验证方法学和库函数系统级性能分析:与原设计相比,优化后二级Cache整体性能约为原设计的2倍。 展开更多
关键词 DSP 二级cachE 存储体 门控脉冲时钟 建立时间 流水线 UVM验证
在线阅读 下载PDF
FlyCache:Recommendation-driven edge caching architecture for full life cycle of video streaming
2
作者 Shaohua Cao Quancheng Zheng +4 位作者 Zijun Zhan Yansheng Yang Huaqi Lv Danyang Zheng Weishan Zhang 《Digital Communications and Networks》 2025年第4期961-973,共13页
With the rapid development of 5G technology,the proportion of video traffic on the Internet is increasing,bringing pressure on the network infrastructure.Edge computing technology provides a feasible solution for opti... With the rapid development of 5G technology,the proportion of video traffic on the Internet is increasing,bringing pressure on the network infrastructure.Edge computing technology provides a feasible solution for optimizing video content distribution.However,the limited edge node cache capacity and dynamic user requests make edge caching more complex.Therefore,we propose a recommendation-driven edge Caching network architecture for the Full life cycle of video streaming(FlyCache)designed to improve users’Quality of Experience(QoE)and reduce backhaul traffic consumption.FlyCache implements intelligent caching management across three key stages:before-playback,during-playback,and after-playback.Specifically,we introduce a cache placement policy for the before-playback stage,a dynamic prefetching and cache admission policy for the during-playback stage,and a progressive cache eviction policy for the after-playback stage.To validate the effectiveness of FlyCache,we developed a user behavior-driven edge caching simulation framework incorporating recommendation mechanisms.Experiments conducted on the MovieLens and synthetic datasets demonstrate that FlyCache outperforms other caching strategies in terms of byte hit rate,backhaul traffic,and delayed startup rate. 展开更多
关键词 Edge caching cache architecture cache placement cache admission caching eviction
在线阅读 下载PDF
Resource Allocation of UAV-Assisted Mobile Edge Computing Systems with Caching
3
作者 Pu Dan Feng Wenjiang Zhang Juntao 《China Communications》 2025年第10期269-279,共11页
In this paper,unmanned aerial vehicle(UAV)is adopted to serve as aerial base station(ABS)and mobile edge computing(MEC)platform for wire-less communication systems.When Internet of Things devices(IoTDs)cannot cope wit... In this paper,unmanned aerial vehicle(UAV)is adopted to serve as aerial base station(ABS)and mobile edge computing(MEC)platform for wire-less communication systems.When Internet of Things devices(IoTDs)cannot cope with computation-intensive and/or time-sensitive tasks,part of tasks is offloaded to the UAV side,and UAV process them with its own computing resources and caching resources.Thus,the burden of IoTDs gets relieved under the satisfaction of the quality of service(QoS)require-ments.However,owing to the limited resources of UAV,the cost of whole system,i.e.,that is defined as the weighted sum of energy consumption and time de-lay with caching,should be further optimized while the objective function and the constraints are non-convex.Therefore,we first jointly optimize commu-nication resources B,computing resources F and of-floading rates X with alternating iteration and convex optimization method,and then determine the value of caching decision Y with branch-and-bound(BB)al-gorithm.Numerical results show that UAV assisting partial task offloading with content caching is supe-rior to local computing and full offloading mechanism without caching,and meanwhile the cost of whole sys-tem gets further optimized with our proposed scheme. 展开更多
关键词 caching MEC resource allocation UAV
在线阅读 下载PDF
Utility-Driven Edge Caching Optimization with Deep Reinforcement Learning under Uncertain Content Popularity
4
作者 Mingoo Kwon Kyeongmin Kim Minseok Song 《Computers, Materials & Continua》 2025年第10期519-537,共19页
Efficient edge caching is essential for maximizing utility in video streaming systems,especially under constraints such as limited storage capacity and dynamically fluctuating content popularity.Utility,defined as the... Efficient edge caching is essential for maximizing utility in video streaming systems,especially under constraints such as limited storage capacity and dynamically fluctuating content popularity.Utility,defined as the benefit obtained per unit of cache bandwidth usage,degrades when static or greedy caching strategies fail to adapt to changing demand patterns.To address this,we propose a deep reinforcement learning(DRL)-based caching framework built upon the proximal policy optimization(PPO)algorithm.Our approach formulates edge caching as a sequential decision-making problem and introduces a reward model that balances cache hit performance and utility by prioritizing high-demand,high-quality content while penalizing degraded quality delivery.We construct a realistic synthetic dataset that captures both temporal variations and shifting content popularity to validate our model.Experimental results demonstrate that our proposed method improves utility by up to 135.9%and achieves an average improvement of 22.6%compared to traditional greedy algorithms and long short-term memory(LSTM)-based prediction models.Moreover,our method consistently performs well across a variety of utility functions,workload distributions,and storage limitations,underscoring its adaptability and robustness in dynamic video caching environments. 展开更多
关键词 Edge caching video-on-demand reinforcement learning utility optimization
在线阅读 下载PDF
A Hierarchical-Based Sequential Caching Scheme in Named Data Networking
5
作者 Zhang Junmin Jin Jihuan +3 位作者 Hou Rui Dong Mianxiong Kaoru Ota Zeng Deze 《China Communications》 2025年第5期48-60,共13页
Named data networking(NDNs)is an idealized deployment of information-centric networking(ICN)that has attracted attention from scientists and scholars worldwide.A distributed in-network caching scheme can efficiently r... Named data networking(NDNs)is an idealized deployment of information-centric networking(ICN)that has attracted attention from scientists and scholars worldwide.A distributed in-network caching scheme can efficiently realize load balancing.However,such a ubiquitous caching approach may cause problems including duplicate caching and low data diversity,thus reducing the caching efficiency of NDN routers.To mitigate these caching problems and improve the NDN caching efficiency,in this paper,a hierarchical-based sequential caching(HSC)scheme is proposed.In this scheme,the NDN routers in the data transmission path are divided into various levels and data with different request frequencies are cached in distinct router levels.The aim is to cache data with high request frequencies in the router that is closest to the content requester to increase the response probability of the nearby data,improve the data caching efficiency of named data networks,shorten the response time,and reduce cache redundancy.Simulation results show that this scheme can effectively improve the cache hit rate(CHR)and reduce the average request delay(ARD)and average route hop(ARH). 展开更多
关键词 hierarchical router named data networking sequential caching
在线阅读 下载PDF
An Efficient Content Caching Strategy for Fog-Enabled Road Side Units in Vehicular Networks
6
作者 Faareh Ahmed Babar Mansoor +1 位作者 Muhammad Awais Javed Abdul Khader Jilani Saudagar 《Computer Modeling in Engineering & Sciences》 2025年第9期3783-3804,共22页
Vehicular networks enable seamless connectivity for exchanging emergency and infotainment content.However,retrieving infotainment data from remote servers often introduces high delays,degrading the Quality of Service(... Vehicular networks enable seamless connectivity for exchanging emergency and infotainment content.However,retrieving infotainment data from remote servers often introduces high delays,degrading the Quality of Service(QoS).To overcome this,caching frequently requested content at fog-enabled Road Side Units(RSUs)reduces communication latency.Yet,the limited caching capacity of RSUs makes it impractical to store all contents with varying sizes and popularity.This research proposes an efficient content caching algorithm that adapts to dynamic vehicular demands on highways to maximize request satisfaction.The scheme is evaluated against Intelligent Content Caching(ICC)and Random Caching(RC).The obtained results show that our proposed scheme entertains more contentrequesting vehicles as compared to ICC and RC,with 33%and 41%more downloaded data in 28%and 35%less amount of time from ICC and RC schemes,respectively. 展开更多
关键词 Vehicular networks fog computing content caching infotainment services
在线阅读 下载PDF
Distributed service caching with deep reinforcement learning for sustainable edge computing in large-scale AI
7
作者 Wei Liu Muhammad Bilal +1 位作者 Yuzhe Shi Xiaolong Xu 《Digital Communications and Networks》 2025年第5期1447-1456,共10页
Increasing reliance on large-scale AI models has led to rising demand for intelligent services.The centralized cloud computing approach has limitations in terms of data transfer efficiency and response time,and as a r... Increasing reliance on large-scale AI models has led to rising demand for intelligent services.The centralized cloud computing approach has limitations in terms of data transfer efficiency and response time,and as a result many service providers have begun to deploy edge servers to cache intelligent services in order to reduce transmission delay and communication energy consumption.However,finding the optimal service caching strategy remains a significant challenge due to the stochastic nature of service requests and the bulky nature of intelligent services.To deal with this,we propose a distributed service caching scheme integrating deep reinforcement learning(DRL)with mobility prediction,which we refer to as DSDM.Specifically,we employ the D3QN(Deep Double Dueling Q-Network)framework to integrate Long Short-Term Memory(LSTM)predicted mobile device locations into the service caching replacement algorithm and adopt the distributed multi-agent approach for learning and training.Experimental results demonstrate that DSDM achieves significant performance improvements in reducing communication energy consumption compared to traditional methods across various scenarios. 展开更多
关键词 Intelligent service Edge caching Deep reinforcement learning Mobility prediction
在线阅读 下载PDF
EdgeAIGC:Model caching and resource allocation for edge artificial intelligence generated content
8
作者 Wu Wen Yibin Huang +3 位作者 Xinxin Zhao Peiying Zhang Kai Liu Guowei Shi 《Digital Communications and Networks》 2025年第6期1941-1950,共10页
With the rapid development of generative artificial intelligence technology,the traditional cloud-based centralized model training and inference face significant limitations due to high transmission latency and costs,... With the rapid development of generative artificial intelligence technology,the traditional cloud-based centralized model training and inference face significant limitations due to high transmission latency and costs,which restrict user-side in-situ Artificial Intelligence Generated Content(AIGC)service requests.To this end,we propose the Edge Artificial Intelligence Generated Content(Edge AIGC)framework,which can effectively address the challenges of cloud computing by implementing in-situ processing of services close to the data source through edge computing.However,AIGC models usually have a large parameter scale and complex computing requirements,which poses a huge challenge to the storage and computing resources of edge devices.This paper focuses on the edge intelligence model caching and resource allocation problems in the Edge AIGC framework,aiming to improve the cache hit rate and resource utilization of edge devices for models by optimizing the model caching strategy and resource allocation scheme,and realize in-situ AIGC service processing.With the optimization objectives of minimizing service request response time and execution cost in resource-constrained environments,we employ the Twin Delayed Deep Deterministic Policy Gradient algorithm for optimization.Experimental results show that,compared with other methods,our model caching and resource allocation strategies can effectively improve the cache hit rate by at least 41.06%and reduce the response cost as well. 展开更多
关键词 Generative AI Edge model caching Resource allocation Edge intelligence
在线阅读 下载PDF
A knowledge graph-based reinforcement learning approach for cooperative caching in MEC-enabled heterogeneous networks
9
作者 Dan Wang Yalu Bai Bin Song 《Digital Communications and Networks》 2025年第4期1236-1244,共9页
Existing wireless networks are flooded with video data transmissions,and the demand for high-speed and low-latency video services continues to surge.This has brought with it challenges to networks in the form of conge... Existing wireless networks are flooded with video data transmissions,and the demand for high-speed and low-latency video services continues to surge.This has brought with it challenges to networks in the form of congestion as well as the need for more resources and more dedicated caching schemes.Recently,Multi-access Edge Computing(MEC)-enabled heterogeneous networks,which leverage edge caches for proximity delivery,have emerged as a promising solution to all of these problems.Designing an effective edge caching scheme is critical to its success,however,in the face of limited resources.We propose a novel Knowledge Graph(KG)-based Dueling Deep Q-Network(KG-DDQN)for cooperative caching in MEC-enabled heterogeneous networks.The KGDDQN scheme leverages a KG to uncover video relations,providing valuable insights into user preferences for the caching scheme.Specifically,the KG guides the selection of related videos as caching candidates(i.e.,actions in the DDQN),thus providing a rich reference for implementing a personalized caching scheme while also improving the decision efficiency of the DDQN.Extensive simulation results validate the convergence effectiveness of the KG-DDQN,and it also outperforms baselines regarding cache hit rate and service delay. 展开更多
关键词 Multi-access edge computing Cooperative caching Resource allocation Knowledge graph Reinforcement learning
在线阅读 下载PDF
Mobility-Aware Edge Caching with Transformer-DQN in D2D-Enabled Heterogeneous Networks
10
作者 Yiming Guo Hongyu Ma 《Computers, Materials & Continua》 2025年第11期3485-3505,共21页
In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic natu... In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic nature of user locations and the potential of device-to-device(D2D)cooperative caching,limiting the reduction of transmission latency.To address this issue,this paper proposes a joint optimization scheme for edge caching that integrates user mobility prediction with deep reinforcement learning.First,a Transformer-based geolocation prediction model is designed,leveraging multi-head attention mechanisms to capture correlations in historical user trajectories for accurate future location prediction.Then,within a three-tier heterogeneous network,we formulate a latency minimization problem under a D2D cooperative caching architecture and develop a mobility-aware Deep Q-Network(DQN)caching strategy.This strategy takes predicted location information as state input and dynamically adjusts the content distribution across small base stations(SBSs)andmobile users(MUs)to reduce end-to-end delay inmulti-hop content retrieval.Simulation results show that the proposed DQN-based method outperforms other baseline strategies across variousmetrics,achieving a 17.2%reduction in transmission delay compared to DQNmethods withoutmobility integration,thus validating the effectiveness of the joint optimization of location prediction and caching decisions. 展开更多
关键词 Mobile edge caching D2D heterogeneous networks deep reinforcement learning transformer model transmission delay optimization
在线阅读 下载PDF
Impact toughness,crack initiation and propagation mechanism of Ti6422 alloy with multi-level lamellar microstructure
11
作者 Jie Shen Zhihao Zhang Jianxin Xie 《International Journal of Minerals,Metallurgy and Materials》 2026年第2期595-609,共15页
The influence of different solution and aging conditions on the microstructure,impact toughness,and crack initiation and propagation mechanisms of the novel α+β titanium alloy Ti6422 was systematically investigated.... The influence of different solution and aging conditions on the microstructure,impact toughness,and crack initiation and propagation mechanisms of the novel α+β titanium alloy Ti6422 was systematically investigated.By adjusting the furnace cooling time after solution treatment and the aging temperature,Ti6422 alloy samples were developed with a multi-level lamellar microstructure,in-cluding microscaleαcolonies and α_(p) lamellae,as well as nanoscale α_(s) phases.Extending the furnace cooling time after solution treatment at 920℃ for 1 h from 240 to 540 min,followed by aging at 600℃ for 6 h,increased the α_(p) lamella content,reduced the α_(s) phase content,expanded theαcolonies and α_(p) lamellae size,and improved the impact toughness from 22.7 to 53.8 J/cm^(2).Additionally,under the same solution treatment,raising the aging temperature from 500 to 700℃ resulted in a decrease in the α_(s) phase content and a growth in the thickness of the α_(p) lamella and α_(s) phase.The impact toughness increased significantly with these changes.Samples with high α_(p) lamellae content or large α_(s) phase size exhibited high crack initiation and propagation energies.Impact deformation caused severe kinking of the α_(p) lamellae in crack initiation and propagation areas,leading to a uniform and high-density kernel average misorientation(KAM)distribu-tion,enhancing plastic deformation coordination and uniformity.Moreover,the multidirectional arrangement of coarserαcolonies and α_(p) lamellae continuously deflect the crack propagation direction,inhibiting crack propagation. 展开更多
关键词 novel titanium alloy multi-level lamellar microstructure impact toughness crack initiation and propagation
在线阅读 下载PDF
基于Memcached缓存的大语言模型的电力工作票风险预警快速响应方法
12
作者 郭凯 程博 +1 位作者 石佳 程磊 《内蒙古电力技术》 2026年第1期61-67,共7页
传统电力工作票风险评估依赖人工判断,易出现风险识别遗漏、评估效率低下等问题。为降低电力作业安全事故风险,构建精准、高效的大语言模型风险预警系统尤为必要。针对大语言模型在电力工作票风险预警中存在的响应速度慢、高并发请求处... 传统电力工作票风险评估依赖人工判断,易出现风险识别遗漏、评估效率低下等问题。为降低电力作业安全事故风险,构建精准、高效的大语言模型风险预警系统尤为必要。针对大语言模型在电力工作票风险预警中存在的响应速度慢、高并发请求处理能力不足的挑战,提出了一种基于分布式内存对象缓存(Memcached)的匹配缓存方案,以优化电力工作票风险预警的运行速度。该方法在服务器端维护高频响应结果缓存,并结合Simhash和余弦相似性算法进行优化,以高效匹配和过滤缓存中的预测结果,从而快速响应用户请求。研究结果表明,该方法能够显著提高电力工作票风险预警的运行速度,有效解决高并发处理的难题,在电力运维领域具有重要的应用价值。 展开更多
关键词 电力工作票 大语言模型 分布式内存对象缓存 风险预警 余弦相似性算法 高并发响应
在线阅读 下载PDF
局域网流媒体Caching代理服务器的实现 被引量:2
13
作者 谭劲 余胜生 周敬利 《计算机科学》 CSCD 北大核心 2003年第10期141-143,共3页
1概述 随着流媒体应用程序在互联网上的广泛应用,必将给Internet的负载带来巨大的变化.基于包交换的Internet不是为实时、不间断的流媒体传输而设计的,因此互联网上的流媒体系统将受到了以下4个方面的限制.
关键词 局域网 流媒体 caching 代理服务器 网络负载
在线阅读 下载PDF
Web Caching技术和CDN技术及其比较分析 被引量:15
14
作者 霍耀森 盛大同 《计算机应用研究》 CSCD 北大核心 2003年第5期135-137,共3页
对为提高服务网络质量而出现的一"拉"一"推"两种技术———WebCaching技术和CDN技术作了简要介绍,并从两者各自常见的体系结构出发对两种技术的性能和特点进行了比较分析,结论是CDN技术在减小源服务器负载等方面优... 对为提高服务网络质量而出现的一"拉"一"推"两种技术———WebCaching技术和CDN技术作了简要介绍,并从两者各自常见的体系结构出发对两种技术的性能和特点进行了比较分析,结论是CDN技术在减小源服务器负载等方面优于单纯的WebCaching技术;最后指出进一步的研究方向。 展开更多
关键词 INTERNET 计算机网络 服务质量 网络带宽 Webcaching技术 CDN技术
在线阅读 下载PDF
GCaching——一种网格协同缓存系统 被引量:3
15
作者 李文中 顾铁成 +2 位作者 李春洪 陆桑璐 陈道蓄 《计算机研究与发展》 EI CSCD 北大核心 2004年第12期2211-2217,共7页
协同缓存通过多个代理缓存服务器的协同工作 ,可以充分利用各服务器的缓存空间 ,提高缓存命中率 网格技术可以方便地共享和整合异构的服务器资源 结合网格和协同技术 ,提出了一种网格协同缓存系统GCaching ,它将地域上分布的多个代理... 协同缓存通过多个代理缓存服务器的协同工作 ,可以充分利用各服务器的缓存空间 ,提高缓存命中率 网格技术可以方便地共享和整合异构的服务器资源 结合网格和协同技术 ,提出了一种网格协同缓存系统GCaching ,它将地域上分布的多个代理缓存服务器组成缓存池 ,充分利用缓存资源 ,协同工作 ,为用户提供更好的服务能力 GCaching系统设计并实现了一种缓存放置和替换算法GCPR ,它使用周期缓存更新策略 ,根据用户访问模式自适应地调整缓存数据的分布 仿真实验表明 。 展开更多
关键词 网格 协同缓存
在线阅读 下载PDF
基于内容的Web Caching 被引量:1
16
作者 张兴军 钱德沛 +2 位作者 刘轶 朱利 李越 《小型微型计算机系统》 CSCD 北大核心 2004年第8期1415-1419,共5页
提出一个新的 Web Caching结构模型—基于内容的 Web Caching.模型综合考虑了 Proxy的操作信息和 Web文档的内容特性 ,界定了虚拟用户团体和 Proxy个性 ,并利用 Ontology技术来刻画 Proxy的个性 ,模拟实验表明 ,结合内容属性可以使得 Web
关键词 Web缓冲 内容特性 知识本体 替换算法
在线阅读 下载PDF
一种基于Sever和Proxy流媒体流行性的Caching策略
17
作者 谭劲 余胜生 周敬利 《计算机科学》 CSCD 北大核心 2003年第4期70-72,101,共4页
1 概述经过了2000、2001两年的社区宽带网建设的高速发展后,摆在中国ISP们面前的任务是如何在已建成的宽带网上开展增值服务,许多ISP尝试在宽带网上开展流媒体(Streaming Media)服务,如视频点播VOD(Video On-Demand)系统。然而,流媒体... 1 概述经过了2000、2001两年的社区宽带网建设的高速发展后,摆在中国ISP们面前的任务是如何在已建成的宽带网上开展增值服务,许多ISP尝试在宽带网上开展流媒体(Streaming Media)服务,如视频点播VOD(Video On-Demand)系统。然而,流媒体对网络带宽和实时性的要求使得流服务器必须能够进行端对端(End-to-End)的拥塞控制和质量调整。 展开更多
关键词 Internet 拥塞控制 代理服务器 流媒体 流行性 caching策略 服务器 宽带网
在线阅读 下载PDF
ASP.NET Caching服务 被引量:3
18
作者 熊姣姣 陈刚 《微机发展》 2005年第1期119-121,124,共4页
WWW上的缓存技术被认为是改进Web性能的有效技术之一,微软的新一代开发语言ASP.NET为开发人员提供了十分完善的缓存服务机制。文章介绍了ASP.NET缓存服务的基本特点以及ASP.NET所提供的几种Caching服务,并通过简单的示例分别说明了每种... WWW上的缓存技术被认为是改进Web性能的有效技术之一,微软的新一代开发语言ASP.NET为开发人员提供了十分完善的缓存服务机制。文章介绍了ASP.NET缓存服务的基本特点以及ASP.NET所提供的几种Caching服务,并通过简单的示例分别说明了每种服务的应用。在Web应用程序中合理应用ASP.NET提供的各种缓存服务,就可以轻而易举地提高Web服务器的访问速度与性能。 展开更多
关键词 缓存 回调 缓存应用程序接口 公共语言运行时间
在线阅读 下载PDF
一种改进的基于分布式Caching的自适应搜索机制 被引量:1
19
作者 李鹏 蔡乐才 肖丹丹 《四川理工学院学报(自然科学版)》 CAS 2008年第3期44-46,50,共4页
文章分析了基于Gnutella协议的非结构P2P网络中利用基于分布式Caching的自适应搜索机制来进行资源搜索与使用统一索引Caching机制相比查询成功率有所降低的问题,提出了两种改进方案。通过实验与统一索引Caching机制比较,改进的搜索机制... 文章分析了基于Gnutella协议的非结构P2P网络中利用基于分布式Caching的自适应搜索机制来进行资源搜索与使用统一索引Caching机制相比查询成功率有所降低的问题,提出了两种改进方案。通过实验与统一索引Caching机制比较,改进的搜索机制在不增加网络流量的条件下,能有效提高查询成功率。 展开更多
关键词 非结构化P2P网络 分布式caching 自适应
在线阅读 下载PDF
一种基于分布式Caching的自适应搜索机制 被引量:1
20
作者 李鹏 蔡乐才 《现代电子技术》 2008年第10期139-141,144,共4页
针对基于Gnutella协议的非结构P2P网络中利用洪泛搜索机制进行资源搜索的网络流量大,效率低等问题,提出一种基于分布式Caching的自适应搜索机制。在使用该搜索机制的Gnutella网络中,所有的节点在逻辑上分成多层,在同一层的节点有相同的g... 针对基于Gnutella协议的非结构P2P网络中利用洪泛搜索机制进行资源搜索的网络流量大,效率低等问题,提出一种基于分布式Caching的自适应搜索机制。在使用该搜索机制的Gnutella网络中,所有的节点在逻辑上分成多层,在同一层的节点有相同的group ID,洪泛查询被限制在与group ID匹配的一层。通过实验与统一索引Caching机制比较,该搜索机制在不增加响应时间的条件下,能有效地减少网络流量,提高搜索效率。 展开更多
关键词 非结构化P2P 分布式caching 自适应搜索 GNUTELLA
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部