Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix NaNs during training on 5-appliances #42

Open
3 of 23 tasks
JackKelly opened this issue Feb 20, 2015 · 0 comments
Open
3 of 23 tasks

Fix NaNs during training on 5-appliances #42

JackKelly opened this issue Feb 20, 2015 · 0 comments
Labels
Milestone

Comments

@JackKelly
Copy link
Owner

Ideas:

  • Find a minimal example which still fails (something that trains fast)
  • Is it:
    • The new nntools code (check by re-running e82)
    • dodgy data (check using old_X and old_y)
    • initialisation?
    • learning rate?
    • different learning algorithm?
    • try concat layer (see this comment)
    • Different cost function.
    • doesn't like producing 5 outputs?
      try single-appliance experts.
    • smaller seq length?
    • smaller batch size?
    • doesn't like range of targets (maybe shouldn't get right to 1?)
    • quantized inputs
    • do we get NaNs if we remove the Conv1D layers?
    • do we get NaNs if we swap from BLSTM to LSTM?
    • do we get NaNs if we remove (B)LSTM layers?
  • Hacks:
    • Save, say, last 5 training examples. When we hit a bad patch, examine these examples.
    • Save network weights. If we get a bad bit of training then revert network to, say, 5 steps ago.
@JackKelly JackKelly added the bug label Feb 20, 2015
@JackKelly JackKelly added this to the ASAP milestone Feb 20, 2015
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant