Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training time too long #6

Open
yuanze-lin opened this issue Dec 18, 2018 · 6 comments
Open

Training time too long #6

yuanze-lin opened this issue Dec 18, 2018 · 6 comments

Comments

@yuanze-lin
Copy link

yuanze-lin commented Dec 18, 2018

I'm trying to reproduce your experimental results on ImageNet, however I find that training resnet networks really takes super long time !!! So I want to know how much time did you use to train resnet on ImageNet dataset??? Thanks !!!

@liujingcs
Copy link
Contributor

It takes about 2 days to prune the ResNet-18 with pruning rate of 0.7. It takes a bit long time. We will improve the efficiency of the algorithm soon.

@yuanze-lin
Copy link
Author

yuanze-lin commented Dec 25, 2018

Are you sure??? Then how do you set the epochs of block-wise and network-wise finetuning respectively??? (on ImageNet dataset)

@liujingcs
Copy link
Contributor

This time does not contain network-wise finetuning. Network-wise finetuning takes about 1 day. We set 20 epochs in block-wise finetuning and 60 epochs in network-wise finetuning. To accelerate the pruning, you can set fewer epochs in block-wise finetuning.

@yuanze-lin
Copy link
Author

I really appreciate for your patient reply. :)

@liujingcs
Copy link
Contributor

It should be noted that we only use subset of imagenet. Therefore, you should set max_sample to 10000.

@Caoliangjie
Copy link

Caoliangjie commented Sep 6, 2019

Thanks for your work. When I want to reproduct the channel pruning example, I found that in DCP/dcp/channel_selections/channel_selections.py line 294, grad_fnorm = cum_grad.mul(cum_grad).sum((2, 3)).sqrt().sum(0) ,some errors occured, it means the 'sum()' function is wrong used. I change it to grad_fnorm = cum_grad.mul(cum_grad).sum(2).sum(2).sqrt().sum(0) then it can work.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants