-
Notifications
You must be signed in to change notification settings - Fork 137
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
train accuracy on svhn dataset doesn't improve #18
Comments
The command you used looks good to me! During our experiments we determined, that it is only possible to create a good model, if you train it with several steps. This means that you first train the model until it is not getting better any more. Then you throw away the recognition part of the model and retrain the model, but you initialize the localization part with your already trained weights and the recognition part with random weights. Hope it helps ;) |
Dear Bartzi, thank you for your answer but it is confuse for me.
Best regards |
|
Hello! I Thank you for your great work. I just started training svhn and it quickly increase training accuracy from 0.25 to 0.5 with 3 epoch, but validation accuracy increased only from 0.240754 to 0.256260. Does it mean that I also need to remove recognition part and restart training?
|
I don't really think that this is the same problem... Did you stop the training after two epochs? |
hi there, at svhn part how you adapt the paths of all images to the path on your machine? |
You could, for instance, do the following:
|
@zolatoi @dmandreev @Bartzi |
Did you check my answers on issue #28? I think we should continue this discussion there. |
Dear Bartzi,
thank you for the great job you did with STN-OCR. I implemented all the steps you described and I launched the learning with train_svhn script but I observe that after 90 epochs the train-accuracy doesn't improve (always between 0.26) and the train-loss is around 2.12. I don't know what happened and how to have better performance. Please find below the command line I used:
python3 train_svhn.py ../datasets/svhn/generated/centered/train.csv ../datasets/svhn/generated/centered/valid.csv --gpus 0,1 --log-dir ./logs --save-model-prefix svhn_train_model -b 100 --lr 1e-5 --zoom 0.5 -ci 500 --char-map ../datasets/svhn/svhn_char_map.json
Best Regards,
The text was updated successfully, but these errors were encountered: