天地一体化信息网络 ›› 2023, Vol. 4 ›› Issue (4): 79-85.doi: 10.11959/j.issn.2096-8930.2023045

• 应用 • 上一篇    

稀疏递归神经网络的可扩展低功耗加速器

金磐石1, 李俊杰2, 王静逸2, 李鹏翀3, 邢磊2, 李晓栋1   

  1. 1 中国建设银行股份有限公司,北京 100034
    2 建信金融科技有限责任公司,上海 321004
    3 浪潮电子信息产业股份有限公司,山东 济南 250000
  • 修回日期:2023-11-30 出版日期:2023-12-01 发布日期:2023-12-01
  • 作者简介:金磐石(1965- ),男,中国建设银行股份有限公司首席信息官,主要从事信息技术系统战略策划、规划、协调和实施工作
    李俊杰(1978- ),男,现就职于建信金融科技有限责任公司,主要从事人工智能推理技术研究工作
    王静逸(1990- ),男,现就职于建信金融科技有限责任公司,主要从事人工智能在金融科技领域的应用研究工作
    李鹏翀(1981- ),男,浪潮电子信息产业股份有限公司网络研发部总经理,主要从事数据中心架构研究工作
    邢磊(1981- ),男,建信金融科技有限责任公司基础技术中心副总裁,主要从事分布式架构的设计研发工作
    李晓栋(1982- ),男,中国建设银行股份有限公司金融科技部技术架构管理处副处长,主要从事技术架构设计工作

Scalable Low Power Accelerator for Sparse Recurrent Neural Network

Panshi JIN1, Junjie LI2, Jingyi WANG2, Pengchong LI3, Lei XING2, Xiaodong LI1   

  1. 1 China Construction Bank Co., Ltd., Beijing 100034, China
    2 Jianxin Financial Technology Co., Ltd., Shanghai 321004, China
    3 Inspur Electronic Information Industry Co., Ltd., Jinan, Shandong 250000, China
  • Revised:2023-11-30 Online:2023-12-01 Published:2023-12-01

摘要:

利用银行网点内边缘计算设备进行客流分析、安全保护、风险防控等应用日益广泛,其中 AI 推理芯片的性能和功耗已经成为边缘计算设备选型的一个非常重要的因素。针对递归神经网络由数据依赖性和低数据重用性导致的功耗大、推理性能弱、能效低,难以在低功耗平台上处理等问题,利用FPGA实现了一种电压可扩展的稀疏循环神经网络(RNN)低功率加速器,并在边缘设计算设备上进行了验证。首先,对稀疏RNN进行分析并采用网络压缩的方法设计了处理阵列;其次,由于稀疏RNN的工作负载不平衡,引入电压缩放方法以保持低功耗和高吞吐量。试验表明,该方法可以显著提高系统的RNN 推理速度并降低芯片的处理功耗。

关键词: RNN, 稀疏, 低功耗, 加速方案

Abstract:

The use of edge computing devices in bank outlets for passenger flow analysis, security protection, risk prevention and control is increasingly widespread, among which the performance and power consumption of AI reasoning chips have become a very important factor in the selection of edge computing devices.Aiming at the problems of recurrent neural network, such as high power consumption, weak reasoning performance and low energy efficiency, which were caused by data dependence and low data reusability, this paper realized a sparse RNN low-power accelerator with scalable voltage by using FPGA, and verifies it on the edge design and calculation equipment.Firstly, the sparse -RNN was analyzed and the processing array was designed by network compression.Secondly, due to the unbalanced workload of sparse RNN, it introduced voltage scaling method to maintain low power consumption and high throughput.Experiments show that this method could significantly improve the RNN reasoning speed of the system and reduce the processing power consumption of the chip.

Key words: RNN, sparse, low power consumption, acceleration scheme

中图分类号: 

No Suggested Reading articles found!