Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts, images, and 🔜 video, up to 5x faster than OpenAI CLIP and LLaVA 🖼️ & 🖋️
-
Updated
Oct 1, 2024 - Python
Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts, images, and 🔜 video, up to 5x faster than OpenAI CLIP and LLaVA 🖼️ & 🖋️
Using Segment-Anything and CLIP to generate pixel-aligned semantic features.
Clipora is a powerful toolkit for fine-tuning OpenCLIP models using Low Rank Adapters (LoRA).
Text-to-image search with OpenCLIP, Docker, Flask, Faiss, etc. and a basic front-end.
Group images by provided labels using OpenAI/CLIP
Using Docker compose to start up Triton with the OpenClip model, to encode text in to vectors
Add a description, image, and links to the openclip topic page so that developers can more easily learn about it.
To associate your repository with the openclip topic, visit your repo's landing page and select "manage topics."