Pruning and Neural Architectures Redesigning for Deep Neural Networks Compression in Mobiles: A Review

Pruning and Neural Architectures Redesigning for Deep Neural Networks Compression in Mobiles: A Review

Ibtihal Ferwana (University of Illinios at Urbana-Champaign, USA), Soumaya Chaffar (University of Prince Mugrin, Saudi Arabia), and Samir Brahim Belhaouari (Hamad Bin Khalifa University, Qatar)
DOI: 10.4018/978-1-6684-3795-7.ch005
OnDemand:
(Individual Chapters)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Mobile applications have been ubiquitous in our daily life. Given the success of Deep Neural Networks (DNNs) in image recognition tasks, DNNs are widely implemented on mobile phone applications. Due to the limited memory and energy on mobile phones, DNNs size and execution time are still roadblocks for efficient processing and instant inferences. Many transformative efforts were able to compress DNNs to the desired size for efficient speed, energy, and memory consumption. In this chapter, two areas of compression: pruning and redesigning efficient neural architectures were discussed. For each, recent advancements and highlight their strengths and limitations are discussed along with showing the improvements brought up by each selected methods and compare them. Comparisons are based on evaluating compression rate, inference time, and accuracy. The aim of this chapter would help practitioners who are implementing DNN based mobile applications to choose a compression approach that satisfies their requirements.
Chapter Preview

Complete Chapter List

Search this Book:
Reset