Queryable
natural-language-image-search
Queryable | natural-language-image-search | |
---|---|---|
5 | 9 | |
2,424 | 927 | |
- | - | |
7.9 | 0.0 | |
19 days ago | over 1 year ago | |
Swift | Jupyter Notebook | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Queryable
-
I accidentally built a meme search engine
You might be interested in this, https://github.com/mazzzystar/Queryable, https://queryable.app/
I run it on my iPhone.
Native app. Doesn't require a network connection (great for privacy).
-
Meta's Segment Anything written with C++ / GGML
I think you would want to use something like CLIP embeddings for image search.
Really enjoyed using this app for iOS: https://github.com/mazzzystar/Queryable
-
Shortcuts ?
This project is open source, so maybe someone will help implement it in the future. :)
-
[P] I open sourced Queryable - a CLIP-based photo search app (SwiftUI)
Many Americans distrust Chinese developers, fearing their photo album privacy would be violated and therefore are reluctant to use the product. I often receive emails from some developers asking about technical details. Now that it's free, why not make the source code available too. The link is: https://github.com/mazzzystar/Queryable.
natural-language-image-search
-
I Built an Image Search Engine Using OpenAI Clip and Images from Wikimedia
You have the option of deleting and posting again I think. Anyways good luck.
Wikimedia search seems to work better for most searches I tried, possibly because of the manual tags etc.
https://imagioo.com/?q=astronaut+with+american+flag
https://commons.wikimedia.org/wiki/Special:MediaSearch?type=...
You might want to include examples where your search is better or just a faq on how to use it.
Nice idea though. It does seem to come in handy when you don't have descriptions of images. Eg: https://github.com/haltakov/natural-language-image-search
-
Simplest way to obtain a network classifying images as Paintings / Not Paintings?
You can give OpenAI's CLIP a shot: https://github.com/openai/CLIP. It's capable of doing zero-shot classification. Here is a neat example of CLIP usage: https://github.com/haltakov/natural-language-image-search.
-
[P] *Semantic* Video Search with OpenAI’s CLIP Neural Network
Does this use the same codebase as https://github.com/haltakov/natural-language-image-search ? Or do you have a different approach?
- Show HN: Search photos using natural language
- Show HN: Search photos on Unsplash using natural language queries
-
OpenAI’s CLIP: Search Images with Descriptions Instead of Keywords
CLIP Image Search
-
Search inside YouTube videos using natural language queries
Yes, this is definitely possible. You can maybe try computing some kind of image distance between frames or some keyframe extraction.
Once you compute the features, the search is very efficient! I tried it for searching in the 2M photos dataset from Unsplash and it takes like 2-3 seconds: https://github.com/haltakov/natural-language-image-search
I plan to run my personal photos through it :)
- Use OpenAI’s CLIP to search 2M photos on Unsplash
What are some alternatives?
clip-retrieval - Easily compute clip embeddings and build a clip retrieval system with them
CLIP - CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
aphantasia - CLIP + FFT/DWT/RGB = text to image/video
natural-language-youtube-search - Search inside YouTube videos using natural language
fastbook - The fastai book, published as Jupyter Notebooks
Awesome-CLIP - Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).
steam-image-search - Search for images on Steam using natural language queries.
MoTIS - [NAACL 2022]Mobile Text-to-Image search powered by multimodal semantic representation models(e.g., OpenAI's CLIP)
datasets - 🎁 5,400,000+ Unsplash images made available for research and machine learning
Puddles - A native SwiftUI app architecture
Chinese-CLIP - Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.