-
|
Hi authors, Thank you very much for putting out an amazing research. I have a question on how network compression without fine-tuning is done. Let's take a model, say resnet18 for example. By my understanding, we first convert resnet18 (pretrained with imagenet) to an INN by py2opt. After that, do we partition parameters into groups? Will each group be approximated by a function which lead to model compression? |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 2 replies
-
|
Hi, thank you for interest in our research. |
Beta Was this translation helpful? Give feedback.
-
|
So dim 0 of the previous layer is always grouped with dim 1 of the following layer. The pruning effect comes from the number of sampling grid points that we choose to compute the integral (the smaller the more prune). When performing the permutation of the weight tensor we are only interested in total variation along dim=0 or dim=1 so in your case we can consider this (16, 1, 5, 5) tensor as tensor of shape (16, 25) and calculate total variation as |
Beta Was this translation helpful? Give feedback.
Hi, thank you for interest in our research.
For structural pruning (resampling) of a neural network we build a list of IntegralGroups. Two parameters belong to the same group if they have a common sampling grid. Example, if you have two sequential convolutional layers
conv_1andconv_2, thenconv_1.weightdim=0,conv_1.biasdim=0,conv_2.weightdim=1 belong to the same group. In residual block of the ResNet that looks likeconv_2(conv_1(x)) + x(I skipped the activation functions) we haveconv_1.weightdim=0,conv_1.biasdim=0,conv_2.weightdim=1 belong to the same group and the other group includesconv_1.weightdim=1 andconv_2.weightdim=0,conv_2.biasdim=0, because input size ofconv_1