Large Language Models (LLMs) like ChatGPT, Claude, and Llama have opened up new attack surfaces despite offering tremendous capabilities.Techniques like the Context Ignoring Attack exploit the limitations in how LLMs process information to potentially bypass safeguards.Prompt Leaking involves trying to extract system prompts to understand model limitations and create targeted attacks.Role Play Attacks leverage the creative scenarios of LLMs to bypass safety measures by engaging the model in unethical roles.Prefix Injection manipulates model responses by adding specific text at the beginning of queries, influencing the output.Refusal Suppression attacks aim to stop LLMs from declining harmful queries by instructing them to avoid refusal statements.Sophisticated attackers combine techniques like refusal suppression and context ignoring for more successful attacks.Understanding vulnerabilities in LLMs is crucial as they become more integrated, leading to an escalating battle between exploiters and defenders.