期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Advanced Design for High-Performance and AI Chips
1
作者 Ying Cao Yuejiao Chen +2 位作者 Xi Fan Hong Fu Bingang Xu 《Nano-Micro Letters》 2026年第1期306-336,共31页
Recent years have witnessed transformative changes brought about by artificial intelligence(AI)techniques with billions of parameters for the realization of high accuracy,proposing high demand for the advanced and AI ... Recent years have witnessed transformative changes brought about by artificial intelligence(AI)techniques with billions of parameters for the realization of high accuracy,proposing high demand for the advanced and AI chip to solve these AI tasks efficiently and powerfully.Rapid progress has been made in the field of advanced chips recently,such as the development of photonic computing,the advancement of the quantum processors,the boost of the biomimetic chips,and so on.Designs tactics of the advanced chips can be conducted with elaborated consideration of materials,algorithms,models,architectures,and so on.Though a few reviews present the development of the chips from their unique aspects,reviews in the view of the latest design for advanced and AI chips are few.Here,the newest development is systematically reviewed in the field of advanced chips.First,background and mechanisms are summarized,and subsequently most important considerations for co-design of the software and hardware are illustrated.Next,strategies are summed up to obtain advanced and AI chips with high excellent performance by taking the important information processing steps into consideration,after which the design thought for the advanced chips in the future is proposed.Finally,some perspectives are put forward. 展开更多
关键词 Artificial intelligence Advanced chips ai chips Design tactics Review and perspective
在线阅读 下载PDF
Robotic computing system and embodied AI evolution:an algorithm-hardware co-design perspective
2
作者 Longke Yan Xin Zhao +7 位作者 Bohan Yang Yongkun Wu Guangnan Dai Jiancong Li Chi-Ying Tsui Kwang-Ting Cheng Yihan Zhang Fengbin Tu 《Journal of Semiconductors》 2025年第10期6-23,共18页
Robotic computing systems play an important role in enabling intelligent robotic tasks through intelligent algo-rithms and supporting hardware.In recent years,the evolution of robotic algorithms indicates a roadmap fr... Robotic computing systems play an important role in enabling intelligent robotic tasks through intelligent algo-rithms and supporting hardware.In recent years,the evolution of robotic algorithms indicates a roadmap from traditional robotics to hierarchical and end-to-end models.This algorithmic advancement poses a critical challenge in achieving balanced system-wide performance.Therefore,algorithm-hardware co-design has emerged as the primary methodology,which ana-lyzes algorithm behaviors on hardware to identify common computational properties.These properties can motivate algo-rithm optimization to reduce computational complexity and hardware innovation from architecture to circuit for high performance and high energy efficiency.We then reviewed recent works on robotic and embodied AI algorithms and computing hard-ware to demonstrate this algorithm-hardware co-design methodology.In the end,we discuss future research opportunities by answering two questions:(1)how to adapt the computing platforms to the rapid evolution of embodied AI algorithms,and(2)how to transform the potential of emerging hardware innovations into end-to-end inference improvements. 展开更多
关键词 robotic computing system embodied ai algorithm-hardware co-design ai chip large-scale ai models
在线阅读 下载PDF
AI Computing Systems for Large Langguage Models Training 被引量:1
3
作者 Zhen-Xing Zhang Yuan-Bo Wen +9 位作者 Han-Qi Lyu Chang Liu Rui Zhang Xia-Qing Li Chao Wang Zi-Dong Du Qi Guo Ling Li Xue-Hai Zhou Yun-Ji Chen 《Journal of Computer Science & Technology》 2025年第1期6-41,共36页
In this paper,we present a comprehensive overview of artificial intelligence(AI)computing systems for large language models(LLMs)training.The rapid advancement of LLMs in recent years,coupled with the widespread adopt... In this paper,we present a comprehensive overview of artificial intelligence(AI)computing systems for large language models(LLMs)training.The rapid advancement of LLMs in recent years,coupled with the widespread adoption of algorithms and applications such as BERT,ChatGPT,and DeepSeek,has sparked significant interest in this field.We classify LLMs into encoder-only,encoder-decoder,and decoder-only models,and briefly analyze their training and inference processes to emphasize their substantial need for computational resources.These operations depend heavily on Alspecific accelerators like GPUs(graphics processing units),TPUs(tensor processing units),and MLUs(machine learning units).However,as the gap widens between the increasing complexity of LLMs and the current capabilities of accelerators,it becomes essential to adopt heterogeneous computing systems optimized for distributed environments to manage the growing computational and memory requirements of LLMs.We delve into the execution and scheduling of LLM algorithms,underlining the critical role of distributed computing strategies,memory management enhancements,and boosting computational efficiency.This paper clarifies the complex relationship between algorithm design,hardware infrastructure,and software optimization,and provides an in-depth understanding of both the software and hardware infrastructure supporting LLMs training,offering insights into the challenges and potential avenues for future development and deployment. 展开更多
关键词 artificial intelligence(ai)chip large language model(LLM) ai computing system ACCELERATOR
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部