Skip to content

GangsuKim/CLIP-pytorch

Repository files navigation

CLIP PyTorch Implementation

Implementation of Learning Transferable Visual Models From Natural Language Supervision
[Paper] [Official Code] [Korean Report]

Architecture of CLIP

CLIP
Source : OpenAI/CLIP

DataSet

Result

  • CLIP-Flick30K
    • Training 211 epoch with Flickr30K
  • CLIP-Flick30K-MSCOCO
    • Use pre-train as Flickr30K 200 epoch checkpoint
    • And train 7 epoch with COCO 2015 Image Captioning Task
DataSet CLIP-Flick30K CLIP-Flickr30K-COCO
Food101 1.1% 1.1%
CIFAR-10 12.2% 16.8%

License

ViT : From lucidrains GitHub repository
GPT-2 : Using OpenAI GPT-2 from transformers library

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages