You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
One possibility is that start_lr is too high. Usually, for CNNs and small dense networks the parameters you use work fine, but for some cases (e.g. RNNs) LR 0.0001 could be too high. Try lower values, e.g. 0.000001.
If it doesn't work even with low starting learning rate, it's possible that there are errors in the implementation of the model.
Try to reduce the dataset to a small size (e.g. 100 examples) and check if the model can learn to predict correct labels on the training set (overfit to the training set).
How to set the values of start_lr, end_lr, batch_size and epochs? In my model, I couldn't get the figures like your example.
I used
lr_finder.find(X_train, y_train, start_lr=0.0001, end_lr=1, batch_size=512, epochs=5)
lr_finder.plot_loss(n_skip_beginning=20, n_skip_end=5)
Then I got a figure like this.
lr_finder.plot_loss_change(sma=20, n_skip_beginning=20, n_skip_end=5, y_lim=(-0.1, 0.1))
Could you give me some advice to use other parameters to find a good lr? Many thanks
The text was updated successfully, but these errors were encountered: