<ul data-eligibleForWebStory="true">Researchers from HiddenLayer have devised a new LLM attack called TokenBreaker.They can bypass certain protections by adding or changing a single character, while the LLM still understands the original intent.The attack targets LLMs using tokenization strategies like Byte Pair Encoding or WordPiece.Tokenization breaks text into tokens for LLMs to process.By adding characters to keywords, protective models can be fooled into thinking prompts are safe.This can bypass defenses to sneak malicious content past filters, potentially leading to malware exposure.The end target can still interpret the manipulated text, rendering the protection model ineffective.Models employing Unigram tokenizers were found to be more resistant to such manipulation.Mitigation strategies include choosing models with stronger tokenization methods.