An Iterative Deep Neural Network Pruning Method Based on Filter Redundancy Assessment

2024 5th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT)(2024)

引用 0|浏览0
暂无评分
摘要
Storage and inference of deep neural network models are resource-intensive, limiting their deployment on edge devices. Structured pruning methods can reduce the resource requirements for model storage and inference by selectively removing low-contributing filters. Current methods do not focus on the functional relations between filters and cannot remove filters that are scored as high-contributing but have redundant feature extraction capabilities. To address this limitation, we propose a structured pruning method, Important filter Decreasing Iteratively (IDI), which removes low-contributing filters and functionally redundant filters in two pruning stages, respectively. The first stage of the IDI removes low contribution filters using a filter contribution evaluation criterion. In the second stage, the Important filter Decreasing (ID) operation removes redundant filters with high functional similarity. Multiple ID operations are performed iteratively to gradually reduce the parameter size of the model without drastically breaking its performance. Experimental results show that compared with existing filter-level pruning methods, the compact model obtained by IDI has higher accuracy. IDI efficiently realizes the compression of deep neural networks.
更多
查看译文
关键词
deep neural network,model compression,structured pruning,similarity evaluation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要