• 中国精品科技期刊
  • CCF推荐A类中文期刊
  • 计算领域高质量科技期刊T1类
Advanced Search
Xiang Taoran, Ye Xiaochun, Li Wenming, Feng Yujing, Tan Xu, Zhang Hao, Fan Dongrui. Accelerating Fully Connected Layers of Sparse Neural Networks with Fine-Grained Dataflow Architectures[J]. Journal of Computer Research and Development, 2019, 56(6): 1192-1204. DOI: 10.7544/issn1000-1239.2019.20190117
Citation: Xiang Taoran, Ye Xiaochun, Li Wenming, Feng Yujing, Tan Xu, Zhang Hao, Fan Dongrui. Accelerating Fully Connected Layers of Sparse Neural Networks with Fine-Grained Dataflow Architectures[J]. Journal of Computer Research and Development, 2019, 56(6): 1192-1204. DOI: 10.7544/issn1000-1239.2019.20190117

Accelerating Fully Connected Layers of Sparse Neural Networks with Fine-Grained Dataflow Architectures

Funds: This work was supported by the National Key Research and Development Plan of China (2018YFB1003501), the National Natural Science Foundation of China (61732018, 61872335, 61802367), the International Partnership Program of Chinese Academy of Sciences (171111KYSB20170032), and the Innovation Project of the State Key Laboratory of Computer Architecture (CARCH3303, CARCH3407, CARCH3502, CARCH3505).
More Information
  • Published Date: May 31, 2019
  • Deep neural network (DNN) is a hot and state-of-the-art algorithm which is widely used in applications such as face recognition, intelligent monitoring, image recognition and text recognition. Because of its high computational complexity, many efficient hardware accelerators have been proposed to exploit high degree of parallel processing for DNN. However, the fully connected layers in DNN have a large number of weight parameters, which imposes high requirements on the bandwidth of the accelerator. In order to reduce the bandwidth pressure of the accelerator, some DNN compression algorithms are proposed. But accelerators which are implemented on FPGAs and ASICs usually sacrifice generality for higher performance and lower power consumption, making it difficult to accelerate sparse neural networks. Other accelerators, such as GPUs, are general enough, but they lead to higher power consumption. Fine-grained dataflow architectures, which break conventional Von Neumann architectures, show natural advantages in processing DNN-like algorithms with high computational efficiency and low power consumption. At the same time, it remains broadly applicable and adaptable. In this paper, we propose a scheme to accelerate the sparse DNN fully connected layers on a hardware accelerator based on fine-grained dataflow architecture. Compared with the original dense fully connected layers, the scheme reduces the peak bandwidth requirement of 2.44×~ 6.17×. In addition, the utilization of the computational resource of the fine-grained dataflow accelerator running the sparse fully-connected layers far exceeds the implementation by other hardware platforms, which is 43.15%, 34.57%, and 44.24% higher than the CPU, GPU, and mGPU, respectively.
  • Related Articles

    [1]He Renhua, Li Bing, Du Yibo, Wang Ying, Li Xiaowei, Han Yinhe. A Survey on Algorithm and Hardware Optimization to LWE-Based Fully Homomorphic Encryption[J]. Journal of Computer Research and Development. DOI: 10.7544/issn1000-1239.202331022
    [2]Bai Lifang, Zhu Yuefei, Li Yongjun, Wang Shuai, Yang Xiaoqi. Research Progress of Fully Homomorphic Encryption[J]. Journal of Computer Research and Development, 2024, 61(12): 3069-3087. DOI: 10.7544/issn1000-1239.202221052
    [3]Zhao Xiufeng, Fu Yu, Song Weitao. Circular Secure Homomorphic Encryption Scheme[J]. Journal of Computer Research and Development, 2020, 57(10): 2117-2124. DOI: 10.7544/issn1000-1239.2020.20200422
    [4]Xu Ming, Fan Yimeng, Jiang Changjun. Time-Varying Underwater Acoustic Channel Based Physical Layer Secret Key Generation Scheme[J]. Journal of Computer Research and Development, 2019, 56(12): 2660-2670. DOI: 10.7544/issn1000-1239.2019.20190040
    [5]Ye Qing, Hu Mingxing, Tang Yongli, Liu Kun, Yan Xixi. Efficient Hierarchical Identity-Based Encryption Scheme from Learning with Errors[J]. Journal of Computer Research and Development, 2017, 54(10): 2193-2204. DOI: 10.7544/issn1000-1239.2017.20170394
    [6]Chen Zhigang, Song Xinxia, Zhao Xiufeng. A Multi-Bit Fully Homomorphic Encryption with Better Key Size from LWE[J]. Journal of Computer Research and Development, 2016, 53(10): 2216-2223. DOI: 10.7544/issn1000-1239.2016.20160431
    [7]Fu Wei, Wu Xiaoping, Ye Qing, Xiao Nong, Lu Xicheng. A Multiple Replica Possession Proving Scheme Based on Public Key Partition[J]. Journal of Computer Research and Development, 2015, 52(7): 1672-1681. DOI: 10.7544/issn1000-1239.2015.20140353
    [8]Liu Mingjie, Wang An. Fully Homomorphic Encryption and Its Applications[J]. Journal of Computer Research and Development, 2014, 51(12): 2593-2603. DOI: 10.7544/issn1000-1239.2014.20131168
    [9]Yang Xiaoyuan, Cai Weiyi, Chen Haibin. Multiple-Authority-Key Functional Encryption: A M-KP-ABE Scheme Based on LMSSS[J]. Journal of Computer Research and Development, 2011, 48(8): 1363-1369.
    [10]Tian Junfeng, Jiao Hongqiang, Li Ning, and Liu Tao. Double Secret Keys and Double Random Numbers Authentication Scheme[J]. Journal of Computer Research and Development, 2008, 45(5): 779-785.
  • Cited by

    Periodical cited type(4)

    1. 张黎,骆春山,谢委员,李蓓蓓. 基于分支混淆算法的隐私数据库自适应加密方法. 计算机与现代化. 2022(03): 43-47 .
    2. 赵武清,王甜,耿新. 基于同态加密的电力业务数据安全研究. 微型电脑应用. 2021(06): 113-116 .
    3. 韩舒艳,努尔买买提·黑力力. 选择性隐藏树型访问结构的CP-ABE方案. 计算机工程. 2020(07): 150-158 .
    4. 刘建华,郑晓坤,郑东,敖章衡. 基于属性加密且支持密文检索的安全云存储系统. 信息网络安全. 2019(07): 50-58 .

    Other cited types(0)

Catalog

    Article views PDF downloads Cited by(4)

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return