A new approach to model compression by merging similar parameter groups within a model has been proposed.The method involves selecting, aligning, and merging separate feed-forward sublayers in Transformer models.The approach has been tested on language modeling, image classification, and machine translation tasks.The results show that the method achieves comparable performance to the original models while removing a significant portion of parameters.