WARP (Weight-space Adaptive Recurrent Prediction) framework unifies weight-space learning with linear recurrence for sequence modeling.
Unlike conventional RNNs, WARP parametrizes the hidden state as the weights of a separate neural network, allowing higher-resolution memory and gradient-free adaptation at test-time.
Empirical validation shows that WARP outperforms state-of-the-art baselines on various classification tasks and offers valuable insights into model's inner workings through weight trajectories.
WARP's efficacy is demonstrated across tasks like sequential image completion, dynamical system reconstruction, and multivariate time series forecasting, showcasing its expressiveness and generalization capabilities.