-
Notifications
You must be signed in to change notification settings - Fork 131
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
reproducing CIFAR10 results for AutoSlim #40
Comments
Hi, Rudy, when I greedy slimming the network, I found that the output_channels of SlimmableConv2d didn't change. Did you encounter the same problem? |
Hi dada, I've actually implemented the AutoSlim myself and cross-referenced this code. I could be wrong but I actually notice some lines of code that I believe to be bugs:
|
Hi, Rudy, Thank you for your reply! I did encounter some problems when running the code at v3.0.0, when I run
and I have set autoslim_resnet_train_val.yml autoslim: True But SlimmableConv2d has no definition of us , so in function get_conv_layers, the length of layers is zero. So it prints Totally 0 layers to slim. Do I need to replace SlimmableConv2d with USConv2d in the network? |
Hi, Rudy |
Hi All, Sorry for the late reply. While I fully understand ImageNet requires more compute which researchers may not have, the results on CIFAR are usually misleading for Neural Architecture Search especially for efficient neural networks. That's part of the reason why I didn't include the CIFAR config in this code. But I can post the configs here for your reference:
|
Please also note that the latest version is released under branch v3.0.0, instead of master branch. |
(I am keeping this issue open and marking it as good first issue) |
Hi, I also encountered the same problem, how did you solve it? |
Hi Jiahui,
Thanks for the great work. I'm trying to reproduce AutoSlim for CIFAR-10 (Table 2).
Could you please provide a detailed hyperparameter you used for it?
I'm able to train the baseline MobileNetV2 1.0x to 7.9 Top-1 error using the following hyperparameters:
To train AutoSlim, I use MobileNetV2 1.5x with the exact same hyperparameter but only trained for 50 epochs on a training set (80% of the real training set). Then, during greedy slimming, I use the extra 20% training set as a validation set to decide channel counts. For greedy slimming, I shrink each layer by a step of 10%, which makes it 10 groups as mentioned in the paper.
The final architecture is trained with the same hyperparameters listed above. But I failed to obtain Top-1 error 6.8% as reported in the paper. I'm getting around 7.8%.
Could you please share with me the final architecture for AutoSlim-MobileNetV2 CIFAR-10 with 88MFLOPs? Also, it would be great if you can let me know the hyperparameters you used for CIFAR experiments.
Thanks,
Rudy
The text was updated successfully, but these errors were encountered: