The article discusses the problem of gradient instability in PyTorch and provides a solution to fix it.
The issue arises when the input features are out of scale, causing one feature to have a more dominant effect on the prediction than others.
One solution is to increase the number of iterations in the training loop, but it may take more time to converge to an optimum minima.
A recommended approach is to normalize the input features using min-max normalization, which scales the features within a comparable range and ensures smoother convergence.