Home > Publications database > LAION-400M: Open Dataset of CLIP-Filtered 400 Million Image-Text Pairs |
Contribution to a conference proceedings | FZJ-2022-00923 |
; ; ; ; ; ; ; ;
2021
Please use a persistent id in citations: http://hdl.handle.net/2128/30478
Abstract: Multi-modal language-vision models trained on hundreds of millions of image-textpairs (e.g. CLIP, DALL-E) gained a recent surge, showing remarkable capability toperform zero- or few-shot learning and transfer even in absence of per-sample labelson target image data. Despite this trend, to date there has been no publicly availabledatasets of sufficient scale for training such models from scratch. To address thisissue, in a community effort we build and release for public LAION-400M, adataset with CLIP-filtered 400 million image-text pairs, their CLIP embeddingsand kNN indices that allow efficient similarity search
![]() |
The record appears in these collections: |