<ul data-eligibleForWebStory="true">Uncertainty quantification in long-form text generation is explored in a new study.The study focuses on uncertainty within sub-groupings of data for large language model outputs.Different methods are used to measure uncertainty at the level of individual claims and across the entire output.Biography generation is used as a test case in this study.Demographic attributes are considered to create subgroups of data.Canonical methods for uncertainty quantification perform well for the entire dataset but struggle with subgroup analysis.Group-conditional methods like multicalibration and multivalid conformal prediction are introduced to address subgroup uncertainties.Additional subgroup information consistently enhances calibration and conformal prediction.The study establishes benchmarks for calibration and conformal prediction in the context of long-form text generation.