Large language models trained on web data may include problematic datapoints.
Complete retraining to eliminate such datapoints is computationally prohibitive.
A new algorithm called MSA proposes an efficient way to estimate and undo the influence of individual datapoints by leveraging model checkpoints.
Experimental results show that MSA outperforms existing machine unlearning algorithms, suggesting it could lead to more flexible large language models capable of data erasure.