Skip to content

Latest commit

 

History

History
11 lines (7 loc) · 598 Bytes

rendered-sst2.md

File metadata and controls

11 lines (7 loc) · 598 Bytes

The Rendered SST2 Dataset

In the paper, we used an image classification dataset called Rendered SST2, to evaluate the model's capability on optical character recognition. To do so, we rendered the sentences in the Standford Sentiment Treebank v2 dataset and used those as the input to the CLIP image encoder.

The following command will download a 131MB archive countaining the images and extract into a subdirectory rendered-sst2:

wget https://openaipublic.azureedge.net/clip/data/rendered-sst2.tgz
tar zxvf rendered-sst2.tgz