How LAION used more data and new ML training techniques to improve image and text embeddings for various applications
By Robert. A Gonsalves
You can see my article on Medium.
The source code and generated images are released under the CC BY-SA license.
- Test CLIP with Library of Congress Images
- Test OpenCLIP with Library of Congress Images
- Index Library of Congress Photos with OpenCLIP
- Search for Library of Congress Photos with OpenCLIP
- Create Captions with CoCa and OpenCLIP
- A. Radford et al., CLIP, Learning Transferable Visual Models From Natural Language Supervision (2021)
- M. Cherti et al., OpenCLIP, Reproducible scaling laws for contrastive language-image learning (2022)
- G. Couairon et al., Embedding Arithmetic of Multimodal Queries for Image Retrieval (2022)
- S. Wang and P. Kanwar, BFloat16: The secret to high performance on Cloud TPUs (2019)
- J. Yu, CoCa: Contrastive Captioners are Image-Text Foundation Models (2022)
To get unlimit
To cite this repository:
@software{GreenLIT,
author = {Gonsalves, Robert A.},
title = {Using OpenCLIP for Image Search and Automatic Captioning},
url = {https://github.com/robgon-art/open-clip},
year = 2023,
month = February
}