期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
An Efficient Multiplier-Less Processing Element on Power-of-2 Dictionary-Based Data Quantization
1
作者 JIAXIANG LI MASAO YANAGISAWA YOUHUA SHI 《Integrated Circuits and Systems》 2024年第1期53-62,共10页
The large-scale neural networks have brought incredible shocks to the world,changing people’s lives and offering vast prospects.However,they also come with enormous demands for computational power and storage pressur... The large-scale neural networks have brought incredible shocks to the world,changing people’s lives and offering vast prospects.However,they also come with enormous demands for computational power and storage pressure,the core of its computational requirements lies in the matrix multiplication units dominated by multiplication operations.To address this issue,we propose an area-power-efficient multiplier-less processing element(PE)design.Prior to implementing the proposed PE,we apply a powerof-2 dictionary-based quantization to the model and effectiveness of this quantization method in preserving the accuracy of the original model is confirmed.In hardware design,we present a standard and one variant‘bi-sign’architecture of the PE.Our evaluation results demonstrate that the systolic array that implement our standard multiplier-less PE achieves approximately 38%lower power-delay-product and 13%smaller core area compared to a conventional multiplication-and-accumulation PE and the bi-sign PE design can even save 37%core area and 38%computation energy.Furthermore,the applied quantization reduces the model size and operand bit-width,leading to decreased on-chip memory usage and energy consumption for memory accesses.Additionally,the hardware schematic facilitates expansion to support other sparsity-aware,energy-efficient techniques. 展开更多
关键词 AI accelerators approximate computing efficient-computing model quantization multiplier-less processing element
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部