典型文献
Design and implementation of near-memory computing array architecture based on shared buffer
文献摘要:
Deep learning algorithms have been widely used in computer vision, natural language process-ing and other fields. However, due to the ever-increasing scale of the deep learning model, the re-quirements for storage and computing performance are getting higher and higher, and the processors based on the von Neumann architecture have gradually exposed significant shortcomings such as con-sumption and long latency. In order to alleviate this problem, large-scale processing systems are shifting from a traditional computing-centric model to a data-centric model. A near-memory compu-ting array architecture based on the shared buffer is proposed in this paper to improve system per-formance, which supports instructions with the characteristics of store-calculation integration, reduc-ing the data movement between the processor and main memory. Through data reuse, the processing speed of the algorithm is further improved. The proposed architecture is verified and tested through the parallel realization of the convolutional neural network ( CNN) algorithm. The experimental re-sults show that at the frequency of 110 MHz, the calculation speed of a single convolution operation is increased by 66 . 64% on average compared with the CNN architecture that performs parallel cal-culations on field programmable gate array( FPGA) . The processing speed of the whole convolution layer is improved by 8 . 81% compared with the reconfigurable array processor that does not support near-memory computing.
文献关键词:
中图分类号:
作者姓名:
SHAN Rui;GAO Xu;FENG Yani;HUI Chao;CUI Xinyue;CHAI Miaomiao
作者机构:
School of Electronic Engineering,Xi'an University of Posts and Telecommunications,Xi'an 710121,P.R.China;School of Computer,Xi'an University of Posts and Telecommunications,Xi'an 710121,P.R.China
文献出处:
引用格式:
[1]SHAN Rui;GAO Xu;FENG Yani;HUI Chao;CUI Xinyue;CHAI Miaomiao-.Design and implementation of near-memory computing array architecture based on shared buffer)[J].高技术通讯(英文版),2022(04):345-353
A类:
B类:
Design,implementation,near,memory,computing,array,architecture,shared,buffer,Deep,learning,algorithms,have,been,widely,used,computer,vision,natural,language,other,fields,However,due,increasing,scale,deep,model,quirements,storage,performance,getting,higher,processors,von,Neumann,gradually,exposed,significant,shortcomings,such,sumption,long,latency,In,order,alleviate,this,problem,large,processing,systems,shifting,from,traditional,centric,data,proposed,paper,which,supports,instructions,characteristics,store,calculation,integration,reduc,movement,between,main,Through,reuse,speed,further,improved,verified,tested,through,parallel,realization,convolutional,neural,network,experimental,sults,show,that,frequency,MHz,single,operation,increased,by,average,compared,performs,culations,programmable,gate,FPGA,whole,layer,reconfigurable,does,not
AB值:
0.527993
相似文献
机标中图分类号,由域田数据科技根据网络公开资料自动分析生成,仅供学习研究参考。