<ul data-eligibleForWebStory="true">Obtaining multiple diverse samples from Large Language Models for a prompt is a challenge.Current methods focus on token-level diversity, leading to repetitive responses.Intent Factored Generation (IFG) proposed to address diversity and engagement issues.IFG involves sampling a semantic intent first and then generating a response based on the intent and prompt.Higher temperature used for intent step to promote diversity, lower temperature for final generation for coherence.Prompting the model to state its intent before generating enhances reasoning tasks.IFG shows effectiveness in improving pass@k and Reinforcement Learning on math and code tasks.IFG combined with Direct Preference Optimization enhances conversational diversity without loss in reward.IFG maintains diversity and quality in general language modeling using reader comments and news articles dataset.IFG is a simple method to boost diversity in Large Language Models while preserving performance.The method is easy to integrate into various algorithms for improved performance across applications.