Training
Example:
Optimum Learning Rate
Refer to this link
What is Learning Rate?
Hyper parameter that decides how much gradient should be back propogated
i.e. How much we move towards minimum
Small learning rate → Converge slowly to minimum
Large learning rate → Diverges

Choose value that is in the middle of the sharpest downward slope
The 1cycle Policy
Refer to this link
Similarly to the regular
learner.fit
, we need to find the optimum learning rate usinglr_finder
.
Last updated
Was this helpful?