期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Topology-Based Data Dissemination Approaches for Large Scale Data Centric Networking Architecture
1
作者 陈佳 张宏科 周华春 《China Communications》 SCIE CSCD 2013年第9期80-96,共17页
Massive information flows are gen- erated from interactive processing and visua- lizations. To efficiently support information transmission over the Interact, information cen- tric architecture has been recently propo... Massive information flows are gen- erated from interactive processing and visua- lizations. To efficiently support information transmission over the Interact, information cen- tric architecture has been recently proposed. In this paper, we consider an information centric architecture, called the data centric networking architecture to provide communication servi- ces to big data, where a service identifier is used to name the data objects. We propose dif- ferent approaches for the dissemination of data objects in a large-scale data centric network. In particular, we propose various approaches to link the data dissemination approach with the topology of the Internet. Further, we eva- luate the proposed approaches with respect to data delivery efficiency, round-trip time imp- rovement, and deployment cost. Based on the results obtained from this study, it can be sh- own that by disseminating data objects to small ISPs, the data delivery efficiency can be significantly improved within an acceptable deployment cost. 展开更多
关键词 data centric networking data dis-semination topology-based
在线阅读 下载PDF
Pragma Directed Shared Memory Centric Optimizations on GPUs 被引量:1
2
作者 Jing Li CCF, Lei Liu +4 位作者 Yuan Wu Xiang-Hua Liu Yi Gao Xiao-Bing Feng Cheng-YongWu 《Journal of Computer Science & Technology》 SCIE EI CSCD 2016年第2期235-252,共18页
GPUs become a ubiquitous choice as coprocessors since they have excellent ability in concurrent processing. In GPU architecture, shared memory plays a very important role in system performance as it can largely improv... GPUs become a ubiquitous choice as coprocessors since they have excellent ability in concurrent processing. In GPU architecture, shared memory plays a very important role in system performance as it can largely improve bandwidth utilization and accelerate memory operations. However, even for affine GPU applications that contain regular access patterns, optimizing for shared memory is not an easy work. It often requires programmer expertise and nontrivial parameter selection. Improper shared memory usage might even underutilize GPU resource: Even using state-of-the-art high level programming models (e.g., OpenACC and OpenHMPP), it is still hard to utilize shared memory since they lack inherent support in describing shared memory optimization and selecting suitable parameters, let alone maintaining high resource utilization. Targeting higher productivity for affine applications, we propose a data centric way to shared memory optimization on GPU. We design a pragma extension on OpenACC so as to convey data management hints of programmers to compiler. Meanwhile, we devise a compiler framework to automatically select optimal parameters for shared arrays, using the polyhedral model. We further propose optimization techniques to expose higher memory and instruction level parallelism. The experimental results show that our shared memory centric approaches effectively improve the performance of five typical GPU applications across four widely used platforms by 3.7x on average, and do not burden programmers with lots of pragmas. 展开更多
关键词 GPU shared memory pragma directed data centric
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部