Tokenization can impact model performance by introducing a phenomenon called tokenization bias.The Byte-Token Representation Lemma establishes a mapping between token and byte-level distributions.A next-byte sampling algorithm eliminates tokenization bias, converting tokenized LMs into token-free ones.The method shows improved performance in fill-in-the-middle tasks and model ensembles across different benchmarks.