Redirigiendo al acceso original de articulo en 22 segundos...
Inicio  /  Applied Sciences  /  Vol: 12 Par: 21 (2022)  /  Artículo
ARTÍCULO
TITULO

Compression of Deep Convolutional Neural Network Using Additional Importance-Weight-Based Filter Pruning Approach

Shrutika S. Sawant    
Marco Wiedmann    
Stephan Göb    
Nina Holzer    
Elmar W. Lang and Theresa Götz    

Resumen

The success of the convolutional neural network (CNN) comes with a tremendous growth of diverse CNN structures, making it hard to deploy on limited-resource platforms. These over-sized models contain a large amount of filters in the convolutional layers, which are responsible for almost 99% of the computation. The key question here arises: Do we really need all those filters? By removing entire filters, the computational cost can be significantly reduced. Hence, in this article, a filter pruning method, a process of discarding a subset of unimportant or weak filters from the original CNN model, is proposed, which alleviates the shortcomings of over-sized CNN architectures at the cost of storage space and time. The proposed filter pruning strategy is adopted to compress the model by assigning additional importance weights to convolutional filters. These additional importance weights help each filter learn its responsibility and contribute more efficiently. We adopted different initialization strategies to learn more about filters from different aspects and prune accordingly. Furthermore, unlike existing pruning approaches, the proposed method uses a predefined error tolerance level instead of the pruning rate. Extensive experiments on two widely used image segmentation datasets: Inria and AIRS, and two widely known CNN models for segmentation: TernausNet and standard U-Net, verify that our pruning approach can efficiently compress CNN models with almost negligible or no loss of accuracy. For instance, our approach could significantly reduce 85% of all floating point operations (FLOPs) from TernausNet on Inria with a negligible drop of 0.32% in validation accuracy. This compressed network is six-times smaller and almost seven-times faster (on a cluster of GPUs) than that of the original TernausNet, while the drop in the accuracy is less than 1%. Moreover, we reduced the FLOPs by 84.34% without significantly deteriorating the output performance on the AIRS dataset for TernausNet. The proposed pruning method effectively reduced the number of FLOPs and parameters of the CNN model, while almost retaining the original accuracy. The compact model can be deployed on any embedded device without any specialized hardware. We show that the performance of the pruned CNN model is very similar to that of the original unpruned CNN model. We also report numerous ablation studies to validate our approach.

 Artículos similares

       
 
Liao He, Qingfeng Li and Baifu An    
Geological hazards, such as the frequent occurrence of rock bursts in deep mining, emphasize the critical necessity for the early warning and prediction of dynamic fractures in coal and rock masses, as well as the destabilization of the surrounding rock.... ver más
Revista: Applied Sciences

 
Hongrui Lu, Yingjun Zhang and Zhuolin Wang    
As the technology for offshore wireless transmission and collaborative innovation in unmanned ships continues to mature, research has been gradually carried out in various countries on methods of compressing and transmitting perceptual video while drivin... ver más

 
Diego Renza and Dora Ballesteros    
CNN models can have millions of parameters, which makes them unattractive for some applications that require fast inference times or small memory footprints. To overcome this problem, one alternative is to identify and remove weights that have a small im... ver más
Revista: Informatics

 
Hu Liu, Meng Li and Yijun Shen    
As oil and natural gas production continue to go deeper into the ocean, the flexible riser, as a connection to the surface of the marine oil and gas channel, will confront greater problems in its practical application. Composite materials are being consi... ver más

 
Jonathan Miquel, Laurent Latorre and Simon Chamaillé-Jammes    
Biologging refers to the use of animal-borne recording devices to study wildlife behavior. In the case of audio recording, such devices generate large amounts of data over several months, and thus require some level of processing automation for the raw d... ver más