Training
Last updated
Last updated
Example:
Refer to this link
Hyper parameter that decides how much gradient should be back propogated
i.e. How much we move towards minimum
Small learning rate → Converge slowly to minimum
Large learning rate → Diverges
Choose value that is in the middle of the sharpest downward slope
Refer to this link
Similarly to the regular learner.fit
, we need to find the optimum learning rate using lr_finder
.