In model compression, filter pruning stands out as a pivotal technique. Its significance becomes particularly crucial as the present deep learning models are developed into larger and more complicated architectures, which leads to massive parameters and high floating-point operations per second (FLOPs). Challenges have appeared due to the high computational demands associated with these advanced model structures. In this work, we introduce two novel methods aimed at addressing the challenges above: innovative automatic filter pruning methods via semi-supervised multi-task learning (SSMTL) hypernetwork and partial weight training hypernetwork, respectively. Both methods effectively train the hypernetwork and enhance the precision of the neural architecture search with reinforcement learning. Compared to other filter pruning methods, our approach achieves higher model accuracy at similar pruning ratios.
|