Researchers have proposed a novel approach for efficient continual learning in binarized neural networks.
The approach integrates stochastically-activated engrams as a gating mechanism for metaplastic binarized neural networks (mBNNs).
This method leverages the computational efficiency of mBNNs combined with the robustness of probabilistic memory traces to mitigate forgetting and maintain model reliability.
The approach achieves high accuracies in class-incremental scenarios, comparable to state-of-the-art methods, while significantly reducing GPU and RAM usage.