![D] Is there a model similar to CLIP but for images only dataset, instead of (image, text) pairs? : r/MachineLearning D] Is there a model similar to CLIP but for images only dataset, instead of (image, text) pairs? : r/MachineLearning](https://preview.redd.it/uck7e6cwo3j81.png?width=1944&format=png&auto=webp&s=19f1d44fbda5516e579568f9407981429f801f72)
D] Is there a model similar to CLIP but for images only dataset, instead of (image, text) pairs? : r/MachineLearning
![CLIP: The Most Influential AI Model From OpenAI — And How To Use It | by Nikos Kafritsas | Towards Data Science CLIP: The Most Influential AI Model From OpenAI — And How To Use It | by Nikos Kafritsas | Towards Data Science](https://miro.medium.com/v2/resize:fit:1400/1*ag6qUFmmXAr4E410Ll-eSQ.png)
CLIP: The Most Influential AI Model From OpenAI — And How To Use It | by Nikos Kafritsas | Towards Data Science
![Aran Komatsuzaki on Twitter: "+ our own CLIP ViT-B/32 model trained on LAION-400M that matches the performance of OpenaI's CLIP ViT-B/32 (as a taste of much bigger CLIP models to come). search Aran Komatsuzaki on Twitter: "+ our own CLIP ViT-B/32 model trained on LAION-400M that matches the performance of OpenaI's CLIP ViT-B/32 (as a taste of much bigger CLIP models to come). search](https://pbs.twimg.com/media/FPNJk5AVIAEsITJ.jpg)
Aran Komatsuzaki on Twitter: "+ our own CLIP ViT-B/32 model trained on LAION-400M that matches the performance of OpenaI's CLIP ViT-B/32 (as a taste of much bigger CLIP models to come). search
![LAION Presents The Largest Freely Available Image-Text Dataset With More Than 5 Billion CLIP-Filtered Image-Text Pairs, 14x Bigger Than LAION-400M - MarkTechPost LAION Presents The Largest Freely Available Image-Text Dataset With More Than 5 Billion CLIP-Filtered Image-Text Pairs, 14x Bigger Than LAION-400M - MarkTechPost](http://www.marktechpost.com/wp-content/uploads/2022/05/Blog-Banner-2.png)