Post-training is important for enhancing reasoning capabilities of large language models.Supervised fine-tuning (SFT) is efficient but may lead to overfitting in larger models.Reinforcement fine-tuning (RFT) generally yields better generalization but depends heavily on base model strength.Unified Fine-Tuning (UFT) combines SFT and RFT into an integrated process, outperforming both methods regardless of model sizes.