menu
techminis

A naukri.com initiative

google-web-stories
Home

>

Robotics News

>

Shielding ...
source image

Unite

1M

read

350

img
dot

Image Credit: Unite

Shielding Prompts from LLM Data Leaks

  • IBM NeurIPS 2024 submission proposes a system to protect users from submitting personal information during conversations with Large Language Models (LLMs).
  • IBM researchers tested user friction with mock-up examples in a study.
  • ChatGPT self-censors responses to critical information prompts but is more tolerant towards personal data.
  • IBM system uses local LLMs or NLP-based heuristics to identify sensitive material in prompts.
  • Proposed system aims to sanitize LLM prompt inputs to prevent data exploitation and privacy breaches.
  • IBM's method intercepts outgoing packets to LLMs at the network level for prompt reformulation.
  • Structured classification allows for pre-defined sensitive attribute identification in prompt reformulation.
  • ChatGPT criticizes third-party intervention in LLM interactions citing potential data mishandling.
  • ChatGPT argues that local LLM might breach user intent and introduce privacy risks.
  • IBM proposal poses user adoption barriers due to implementation complexities according to ChatGPT.

Read Full Article

like

21 Likes

For uninterrupted reading, download the app