MLX-VLM is a package for running Vision LLMs locally on your Mac using MLX.
-
Updated
Jun 22, 2024 - Python
MLX-VLM is a package for running Vision LLMs locally on your Mac using MLX.
OpenMMLab Detection Toolbox and Benchmark
Optimized local inference for LLMs with HuggingFace-like APIs for quantization, vision/language models, multimodal agents, speech, vector DB, and RAG.
Scenic: A Jax Library for Computer Vision Research and Beyond
CAIRI Supervised, Semi- and Self-Supervised Visual Representation Learning Toolbox and Benchmark
MIST: A simple, scalable, and end-to-end framework for 3D medical imaging segmentation.
Extract markdown and images from URLs, PDFs, docs, slides, and more, ready for multimodal LLMs. ⚡
Real Time Detection Transformer and Tracking
A Deep Learning Approach to Object Detection and Tracking: Vision Transformers and DeepSORT
Parameter Efficient Fine-tuning of Self-supervised ViTs without Catastrophic Forgetting
Open source implementation of "Vision Transformers Need Registers"
Omni Geoguessr AI: A Vision Transformer AI integrated with Geoguessr for automated geographic location prediction and gameplay using streetview panoramas.
An introduction to attention mechanisms and the vision transformer
Simplified Pytorch implementation of Vision Transformer (ViT) for small datasets like MNIST, FashionMNIST, SVHN and CIFAR10.
Source code for paper: https://arxiv.org/pdf/2305.17370.pdf
The code repository for "Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning"(CVPR24) in PyTorch.
InternLM-XComposer2 is a groundbreaking vision-language large model (VLLM) excelling in free-form text-image composition and comprehension.
Official PyTorch implementation of the CVPR 2024 paper: State Space Models for Event Cameras (Spotlight).
[CVPR 2024] Code for our Paper "DeiT-LT: Distillation Strikes Back for Vision Transformer training on Long-Tailed Datasets"
Add a description, image, and links to the vision-transformer topic page so that developers can more easily learn about it.
To associate your repository with the vision-transformer topic, visit your repo's landing page and select "manage topics."