-
Notifications
You must be signed in to change notification settings - Fork 130
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Have you tested the codes with other models? #7
Comments
Hi, @shuxjweb. Did you trained on SVHN dataset? I have trained on SVHN dataset with same experiment setting as the paper, but there is always 1% point accuracy lower than paper's result. |
@blankWorld I only tested it on cifar10, and the accuracy in paper also has a fluctuate range, see B.1 and B.2. |
have u try another models? like res2net? resnxet? |
@ClayZhang0403 I just tested the codes with wideresnet and resnet50, but resnet50 behaved much worse than wideresnet. |
different network may need different parameters, such as batch size and lr |
Could you share us with the hyper-parameters? I only get 86% acc with 250 labels by the default hyper-parameters. |
Has anybody tried replicating for CIFAR 100 using this code? I get 6-7% less accuracy than the reported numbers for n=10000 labelled samples |
cifar100 need larger model, in this paper, they use 135 filters per layer in wideresnet when training cifar100... |
Hi, has anyone tried with larger model with 1 millions of params? I found that there is a significant decrease in accuracy when I ran with larger models. model: WRN_28_2
model: wide_resnet50_2
model: vgg19
|
Firstly, thanks for you great work.
I have done several experiments using your codes, and most experimental results are satisfied. For cifar10, the best acc is 0.8982 with 250 labels, 0.9438 with 4000labels. However, these results are based on the model of wideresnet, which is utilized in paper. The accuracy woud be much worse when I alternate it with resnet50. The best acc is 0.7384 with 250 labels, 0.8219 with 4000labels. I wonder why different models produce such different results ?
The text was updated successfully, but these errors were encountered: