Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Depth loss goes negative during long training run #13

Open
gch opened this issue Aug 8, 2022 · 4 comments
Open

Depth loss goes negative during long training run #13

gch opened this issue Aug 8, 2022 · 4 comments

Comments

@gch
Copy link

gch commented Aug 8, 2022

To try to reproduce the paper results (like other bug reporters, I get worse results on a stock training compared to authors), I left my training running far past the typical 200 epochs. At around ~325 epochs, I found that the loss_depth value went negative. I'm assuming this is an error. Have you observed this in practice?

I will spend some time probing into why this is happening and update this ticket as needed.

@2gunsu
Copy link

2gunsu commented Aug 13, 2022

Hello, I'm not the author of this paper, but may I ask if the performance improved when training with longer epochs?
I have the same problem that the performance of this paper is not reproduced.

@gch
Copy link
Author

gch commented Aug 14, 2022

Performance kept improving, minus this oddity where the depth loss went negative (which I still need to debug). But, for example, my loss for car easy 3D @ 0.7 was 23.8, but the reported performance for the pretrained model is more like 26.

There's definitely general concern with this repo in the ability to reproduce the pretrained model reliably. Not sure if it simply takes a bunch of retrying to get the best performance?

@2gunsu
Copy link

2gunsu commented Aug 15, 2022

Thank you for your reply.
I am also making many attempts to reproduce, and if I find a solution, I will share it.

@FlyingAnt2018
Copy link

Performance kept improving, minus this oddity where the depth loss went negative (which I still need to debug). But, for example, my loss for car easy 3D @ 0.7 was 23.8, but the reported performance for the pretrained model is more like 26.

There's definitely general concern with this repo in the ability to reproduce the pretrained model reliably. Not sure if it simply takes a bunch of retrying to get the best performance?

cfg.SEED = 1903919922 can reproduce official result

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants