menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Bridging t...
source image

Arxiv

1w

read

296

img
dot

Image Credit: Arxiv

Bridging the Gap Between Preference Alignment and Machine Unlearning

  • Mainstream methods like Reinforcement Learning with Human Feedback (RLHF) face challenges in Preference Alignment (PA) for Large Language Models (LLMs).
  • High-quality datasets of positive preference examples are costly and computationally intensive, limiting their use in low-resource scenarios.
  • LLM unlearning technique presents a promising alternative by directly removing the influence of negative examples.
  • A framework called Unlearning to Align (U2A) is proposed to optimize the selection and unlearning of negative examples for improved PA performance.

Read Full Article

like

17 Likes

For uninterrupted reading, download the app