Queryable
The open-source code of Queryable, an iOS app, utilizes the CLIP model to conduct offline searches in the Photos
album.
Unlike the object recognition-based search feature built into the iOS gallery, Queryable allows you to use natural language statements, such as a brown dog sitting on a bench
, to search your gallery. It operates offline, ensuring that your album privacy won't be leaked to anyone, including Apple/Google.
Performance
final.mp4
How does it work?
- Process all photos in your album through the CLIP Image Encoder to create a set of local image vectors.
- When a new text query is inputted, convert the text into a text vector using the Text Encoder.
- Compare the text vector with all the stored image vectors, evaluating the level of similarity between the text query and each image.
- Sort and return the top K most similar results.
The process is as follows:
For more details, please refer to my article Run CLIP on iPhone to Search Photos.
Run on Xcode
Download the ImageEncoder_float32.mlmodelc
and TextEncoder_float32.mlmodelc
from Google Drive.
Clone this repo, put the downloaded models below CoreMLModels/
path and run Xcode, it should work.
Contributions
Disclaimer: I am not a professional iOS engineer, please forgive my poor Swift code. You may focus only on the loading, computation, storage, and sorting of the model.
You can apply Queryable to your own business product, but I don't recommend modifying the appearance directly and then listing it on the App Store. If you are interested in optimizing certain aspects, feel free to submit a PR (Pull Request).
If you have any questions/suggestions, here are some contact methods: Discord | Twitter | Reddit: r/Queryable.
License
MIT License
Copyright (c) 2023 Ke Fang