menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Preserving...
source image

Arxiv

1w

read

245

img
dot

Image Credit: Arxiv

Preserving Diversity in Supervised Fine-Tuning of Large Language Models

  • Large Language Models (LLMs) typically rely on Supervised Fine-Tuning (SFT) to specialize in downstream tasks.
  • Cross Entropy (CE) loss leads to reduced diversity in the model's outputs, hindering further development.
  • A new game-theoretic formulation for SFT is introduced to address output diversity limitations.
  • The proposed approach enhances output diversity without compromising downstream performance.

Read Full Article

like

14 Likes

For uninterrupted reading, download the app