BLIP3-KALE is an open-source dataset of 218 million image-text pairs.It addresses the limitations of previous image caption datasets by combining synthetic captions and real-world information.KALE features knowledge-augmented dense captions that provide rich descriptive detail and factual grounding.It sets a new benchmark for density and factual grounding in image descriptions.