Large Language Models (LLMs) typically rely on Supervised Fine-Tuning (SFT) to specialize in downstream tasks.Cross Entropy (CE) loss leads to reduced diversity in the model's outputs, hindering further development.A new game-theoretic formulation for SFT is introduced to address output diversity limitations.The proposed approach enhances output diversity without compromising downstream performance.