Researchers have explored the dynamics of artificial neural network trajectory during training with unconventional large learning rates.
For certain learning rate values, the optimization shifts from exploitation-like to exploration-exploitation balance, leading to sensitive dependence on initial conditions.
Training time to achieve acceptable accuracy in the test set reduces to a minimum in this regime, indicating accelerated training of neural networks near the onset of chaos.
The study, initially demonstrated on the MNIST classification task, shows the constructive role of transient chaotic dynamics in training artificial neural networks across various learning tasks and architectures.