5.85 billion CLIP-filtered image-text pairs — the largest open multimodal dataset.
Browse commercial Multimodal → Visit original source ↗LAION-5B is the largest openly-accessible image-text dataset, comprising 5.85 billion image-text pairs scraped from Common Crawl and CLIP-filtered for quality. It powered training of Stable Diffusion and many other open-source multimodal models. The dataset is split by language (2.32B English pairs, 2.26B multi-lingual, 1.27B unknown). Note: LAION provides URLs and metadata, not the images themselves.
LQS is our 7-dimension quality score, computed from the dataset's published statistics. See methodology →
Composite score computed from the 7 dimensions below: completeness, uniqueness, validation health, size adequacy, format compliance, label density, and class balance.
Common tasks and benchmarks where LAION-5B is the default or competitive choice.
What's actually in the dataset — from the maintainer's published stats.
LAION-5B is distributed under CC BY 4.0 (metadata only). This is a third-party public dataset; LabelSets indexes and scores it but does not host or redistribute the data. Always verify current license terms with the maintainer before commercial use.
LabelSets sellers offer paid multimodal datasets with what public datasets often can't give you:
Other entries in the Multimodal catalog.