The rise of Large Language Models (LLMs) has revolutionized artificial intelligence, but they are vulnerable to adversarial perturbations.
The Robustness Measurement and Assessment (RoMA) framework is adapted to quantify LLM resilience against adversarial inputs.
RoMA's estimates demonstrate accuracy with minimal error margins and computational efficiency.
LLM robustness varies between models, categories within the same task, and types of perturbations, highlighting the need for task-specific evaluations.