This paper introduces a new side-channel in large language models (LLMs) that allows an adversary to extract sensitive information about inference inputs.
The side-channel is based on the number of output tokens in the LLM response.
The paper demonstrates attacks utilizing this side-channel in machine translation tasks and text classification tasks.
Proposed mitigations against the output token count side-channel are also discussed.