Loading...
Search for: memory-bandwidths
0.004 seconds

    Partition pruning: Parallelization-aware pruning for dense neural networks

    , Article 28th Euromicro International Conference on Parallel, Distributed and Network-Based Processing, PDP 2020, 11 March 2020 through 13 March 2020 ; 2020 , Pages 307-311 Shahhosseini, S ; Albaqsami, A ; Jasemi, M ; Bagherzadeh, N ; Sharif University of Technology
    Institute of Electrical and Electronics Engineers Inc  2020
    Abstract
    As recent neural networks are being improved to be more accurate, their model's size is exponentially growing. Thus, a huge number of parameters requires to be loaded and stored from/in memory hierarchy and computed in processors to perform training or inference phase of neural network processing. Increasing the number of parameters causes a big challenge for real-time deployment since the memory bandwidth improvement's trend cannot keep up with models' complexity growing trend. Although some operations in neural networks processing are computational intensive such as convolutional layer computing, computing dense layers face with memory bandwidth bottleneck. To address the issue, the paper...