IBM NeurIPS 2024 submission proposes a system to protect users from submitting personal information during conversations with Large Language Models (LLMs).
IBM researchers tested user friction with mock-up examples in a study.
ChatGPT self-censors responses to critical information prompts but is more tolerant towards personal data.
IBM system uses local LLMs or NLP-based heuristics to identify sensitive material in prompts.
Proposed system aims to sanitize LLM prompt inputs to prevent data exploitation and privacy breaches.
IBM's method intercepts outgoing packets to LLMs at the network level for prompt reformulation.
Structured classification allows for pre-defined sensitive attribute identification in prompt reformulation.
ChatGPT criticizes third-party intervention in LLM interactions citing potential data mishandling.
ChatGPT argues that local LLM might breach user intent and introduce privacy risks.
IBM proposal poses user adoption barriers due to implementation complexities according to ChatGPT.