期刊文献+
共找到4篇文章
< 1 >
每页显示 20 50 100
QoS Constrained Network Coding Technique to Data Transmission Using IoT
1
作者 A.Sathishkumar T.Rammohan +5 位作者 S.Sathish Kumar J.Uma K.Srujan Raju Aarti Sangwan M.Sivachitra M.Prabu 《Computer Systems Science & Engineering》 SCIE EI 2022年第11期531-544,共14页
The research work presents,constrained network coding technique to ensure the successful data transmission based composite channel cmos technology using dielectric properties.The charge fragmentation and charge splitt... The research work presents,constrained network coding technique to ensure the successful data transmission based composite channel cmos technology using dielectric properties.The charge fragmentation and charge splitting are two components of the filtered switch domino(FSD)technique.Further behavior of selected switching is achieved using generator called conditional pulse generator which is employed in Multi Dynamic Node Domino(MDND)technique.Both FSD and MDND technique need wide area compared to existing single nodekeeper domino technique.The aim of this research is to minimize dissipation of power and to achieve less consumption of power.The proposed research,works by introducing the method namely Interference and throughput aware Optimized Multicast Routing Protocol(IT-OMRP).The main goal of this proposed research method is to introduce the system which can forward the data packets towards the destination securely and successfully.To achieve the bandwidth and throughput in optimized data transmission,proposed multicast tree is selected by Particle Swarm Optimization which will select the most optimal host node as the branches of multi cast tree.Here node selection is done by considering the objectives residual energy,residual bandwidth and throughput.After node selection multi cast routing is done with the concern of interference to ensure the reliable and successful data transmission.In case of transmission range size is higher than the coverage sense range,successful routing is ensured by selecting secondary host forwarders as a backup which will act as intermediate relay forwarders.The NS2 simulator is used to evaluate research outcome from which it is proved that the proposed technique tends to have increased packet delivery ratio than the existing work. 展开更多
关键词 Multicast routing optimal node selection secondary relay nodes probability of interference residual energy BANDWIDTH THROUGHPUT
在线阅读 下载PDF
Knowledge-Enhanced Bilingual Textual Representations for Cross-Lingual Semantic Textual Similarity
2
作者 Hsuehkuan Lu Yixin Cao +1 位作者 Hou Lei Juanzi Li 《国际计算机前沿大会会议论文集》 2019年第1期436-440,共5页
Joint learning of words and entities is advantageous to various NLP tasks, while most of the works focus on single language setting. Cross-lingual representations learning receives high attention recently, but is stil... Joint learning of words and entities is advantageous to various NLP tasks, while most of the works focus on single language setting. Cross-lingual representations learning receives high attention recently, but is still restricted by the availability of parallel data. In this paper, a method is proposed to jointly embed texts and entities on comparable data. In addition to evaluate on public semantic textual similarity datasets, a task (cross-lingual text extraction) was proposed to assess the similarities between texts and contribute to this dataset. It shows that the proposed method outperforms cross-lingual representations methods using parallel data on cross-lingual tasks, and achieves competitive results on mono-lingual tasks. 展开更多
关键词 Text and knowledge REPRESENTATIONS Cross-lingual REPRESENTATIONS Cross-lingual SEMANTIC TEXTUAL SIMILARITY
在线阅读 下载PDF
Convergence-aware operator-wise mixed-precision training
3
作者 Wenhao Dai Ziyi Jia +1 位作者 Yuesi Bai Qingxiao Sun 《CCF Transactions on High Performance Computing》 2025年第1期43-57,共15页
With the support of more precision formats in emerging hardware architectures,mixed-precision has become a popular approach to accelerate deep learning(DL)training.Applying low-precision formats such as FP16 and BF16 ... With the support of more precision formats in emerging hardware architectures,mixed-precision has become a popular approach to accelerate deep learning(DL)training.Applying low-precision formats such as FP16 and BF16 to neural operators can save GPU memory while improving bandwidth.However,DL frameworks use black and white lists as default mixed-precision selections and cannot flexibly adapt to a variety of neural networks.In addition,existing work on automatic precision adjustment does not consider model convergence,and the decision cost of precision selection is high.To address the above problems,this paper proposes CoMP,a non-intrusive framework for Convergence-aware operator-wise Mixedprecision training.CoMP uses two-stage precision adjustment based on epochs and batches to ensure convergence and performance respectively.After that,CoMP performs subsequent training according to the searched optimal operator-wise mixed-precision plan.The experimental results on A100 GPU show that CoMP achieves a maximum performance speedup of 1.15×compared with PyTorch AMP implementation,while also saving up to 29.81%of GPU memory. 展开更多
关键词 GPU Mixed-precision Neural network training AUTO-TUNING Performance optimization
在线阅读 下载PDF
_(ν)GNN:Non‑Uniformly partitioned full‑graph GNN training on mixed GPUs
4
作者 Hemeng Wang Wenqing Lin +1 位作者 Qingxiao Sun Weifeng Liu 《CCF Transactions on High Performance Computing》 2025年第4期305-322,共18页
Graph neural networks(GNNs)can be adapted to GPUs with high computing capability due to massive arithmetic opera-tions.Compared with mini-batch training,full-graph training does not require sampling of the input graph... Graph neural networks(GNNs)can be adapted to GPUs with high computing capability due to massive arithmetic opera-tions.Compared with mini-batch training,full-graph training does not require sampling of the input graph and halo region,avoiding potential accuracy losses.Current deep learning frameworks evenly partition large graphs to scale GNN training to distributed multi-GPU platforms.On the other hand,the rapid revolution of hardware requires technology companies and research institutions to frequently update their equipment to cope with the latest tasks.This results in a large-scale cluster with a mixture of GPUs with various computational capabilities and hardware specifications.However,existing works fail to consider sub-graphs adapted to different GPU generations,leading to inefficient resource utilization and degraded training efficiency.Therefore,we propose_(ν)GNN,a Non-Uniformly partitioned full-graph GNN training framework on heterogeneous distributed platforms._(ν)GNN first models the GNN processing ability of hardware based on various theoretical parameters.Then,_(ν)GNN automatically obtains a reasonable task partitioning scheme by combining hardware,model,and graph dataset information.Finally,_(ν)GNN implements an irregular graph partitioning mechanism that allows GNN training tasks to execute efficiently on distributed heterogeneous systems.The experimental results show that in real-world scenarios with a mixture of GPU generations,_(ν)GNN can outperform other static partitioning schemes based on hardware specifications. 展开更多
关键词 Graph neural network Distributed training Graph partitioning GPU
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部