Paper ID | SS-NNC.5 | ||
Paper Title | ON THE ROLE OF STRUCTURED PRUNING FOR NEURAL NETWORK COMPRESSION | ||
Authors | Andrea Bragagnolo, Enzo Tartaglione, Attilio Fiandrotti, Marco Grangetto, University of Turin, Italy | ||
Session | SS-NNC: Special Session: Neural Network Compression and Compact Deep Features | ||
Location | Area B | ||
Session Time: | Tuesday, 21 September, 08:00 - 09:30 | ||
Presentation Time: | Tuesday, 21 September, 08:00 - 09:30 | ||
Presentation | Poster | ||
Topic | Special Sessions: Neural Network Compression and Compact Deep Features: From Methods to Standards | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | This works explores the benefits of structured parameter pruning in the framework of the MPEG standardization efforts for neural network compression. First less relevant parameters are pruned from the network, then remaining parameters are quantized and finally quantized parameters are entropy coded. We consider an unstructured pruning strategy that maximizes the number of pruned parameters at the price of randomly sparse tensors and a structured strategy that prunes fewer parameters yet yields regularly sparse tensors. We show that structured pruning enables better end-to-end compression despite lower pruning ratio because it boosts the efficiency of the arithmetic coder. As a bonus, once decompressed, the network memory footprint is lower as well as its inference time. |