menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

NoProp: Tr...
source image

Arxiv

1d

read

254

img
dot

Image Credit: Arxiv

NoProp: Training Neural Networks without Back-propagation or Forward-propagation

  • The paper introduces a new learning method named NoProp, which does not rely on either forward or backward propagation in deep learning.
  • NoProp takes inspiration from diffusion and flow matching methods to independently learn to denoise a noisy target at each layer.
  • The method demonstrates superior accuracy, ease of use, and computational efficiency compared to other back-propagation-free methods on image classification benchmarks such as MNIST, CIFAR-10, and CIFAR-100.
  • NoProp alters the traditional gradient-based learning paradigm, enabling more efficient distributed learning and potentially impacting other characteristics of the learning process.

Read Full Article

like

15 Likes

For uninterrupted reading, download the app