首站-论文投稿智能助手
典型文献
Design and implementation of near-memory computing array architecture based on shared buffer
文献摘要:
Deep learning algorithms have been widely used in computer vision, natural language process-ing and other fields. However, due to the ever-increasing scale of the deep learning model, the re-quirements for storage and computing performance are getting higher and higher, and the processors based on the von Neumann architecture have gradually exposed significant shortcomings such as con-sumption and long latency. In order to alleviate this problem, large-scale processing systems are shifting from a traditional computing-centric model to a data-centric model. A near-memory compu-ting array architecture based on the shared buffer is proposed in this paper to improve system per-formance, which supports instructions with the characteristics of store-calculation integration, reduc-ing the data movement between the processor and main memory. Through data reuse, the processing speed of the algorithm is further improved. The proposed architecture is verified and tested through the parallel realization of the convolutional neural network ( CNN) algorithm. The experimental re-sults show that at the frequency of 110 MHz, the calculation speed of a single convolution operation is increased by 66 . 64% on average compared with the CNN architecture that performs parallel cal-culations on field programmable gate array( FPGA) . The processing speed of the whole convolution layer is improved by 8 . 81% compared with the reconfigurable array processor that does not support near-memory computing.
文献关键词:
作者姓名:
SHAN Rui;GAO Xu;FENG Yani;HUI Chao;CUI Xinyue;CHAI Miaomiao
作者机构:
School of Electronic Engineering,Xi'an University of Posts and Telecommunications,Xi'an 710121,P.R.China;School of Computer,Xi'an University of Posts and Telecommunications,Xi'an 710121,P.R.China
引用格式:
[1]SHAN Rui;GAO Xu;FENG Yani;HUI Chao;CUI Xinyue;CHAI Miaomiao-.Design and implementation of near-memory computing array architecture based on shared buffer)[J].高技术通讯(英文版),2022(04):345-353
A类:
B类:
Design,implementation,near,memory,computing,array,architecture,shared,buffer,Deep,learning,algorithms,have,been,widely,used,computer,vision,natural,language,other,fields,However,due,increasing,scale,deep,model,quirements,storage,performance,getting,higher,processors,von,Neumann,gradually,exposed,significant,shortcomings,such,sumption,long,latency,In,order,alleviate,this,problem,large,processing,systems,shifting,from,traditional,centric,data,proposed,paper,which,supports,instructions,characteristics,store,calculation,integration,reduc,movement,between,main,Through,reuse,speed,further,improved,verified,tested,through,parallel,realization,convolutional,neural,network,experimental,sults,show,that,frequency,MHz,single,operation,increased,by,average,compared,performs,culations,programmable,gate,FPGA,whole,layer,reconfigurable,does,not
AB值:
0.527993
相似文献
Toward memristive in-memory computing:principles and applications
Han Bao;Houji Zhou;Jiancong Li;Huaizhi Pei;Jing Tian;Ling Yang;Shengguang Ren;Shaoqin Tong;Yi Li;Yuhui He;Jia Chen;Yimao Cai;Huaqiang Wu;Qi Liu;Qing Wan;Xiangshui Miao-School of Integrated Circuits,School of Optical and Electronic Information,Wuhan National Laboratory for Optoelectronics,Optics Valley Laboratory,Huazhong University of Science and Technology,Wuhan 430074,China;Hubei Yangtze Memory Laboratories,Wuhan 430205,China;AI Chip Center for Emerging Smart Systems,InnoHK Centers,Hong Kong Science Park,Hong Kong,China;School of Integrated Circuits,Peking University,Beijing 100871,China;School of Integrated Circuits,Beijing National Research Center for Information Science and Technology(BNRist),Tsinghua University,Beijing 100084,China;Frontier Institute of Chip and System,Fudan University,Shanghai 200433,China;School of Electronic Science and Engineering,and Collaborative Innovation Centre of Advanced Microstructures,Nanjing University,Nanjing 210093,China
机标中图分类号,由域田数据科技根据网络公开资料自动分析生成,仅供学习研究参考。