You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There are 2 models in Clip, the image processing model and a text embedding model. Elastic does not support image processing models but if you want to use the text embedding model you can install the Sentence Transformers implementation: https://huggingface.co/sentence-transformers/clip-ViT-B-32-multilingual-v1
2024-02-10 16:04:17,677 INFO : Establishing connection to Elasticsearch
2024-02-10 16:04:17,728 INFO : Connected to cluster named 'docker-cluster' (version: 8.12.0)
2024-02-10 16:04:17,729 INFO : Loading HuggingFace transformer tokenizer and model 'openai/clip-vit-base-patch32'
Traceback (most recent call last):
File "/Users/robinyang/Library/Python/3.9/bin/eland_import_hub_model", line 8, in
sys.exit(main())
File "/Users/robinyang/Library/Python/3.9/lib/python/site-packages/eland/cli/eland_import_hub_model.py", line 254, in main
tm = TransformerModel(
File "/Users/robinyang/Library/Python/3.9/lib/python/site-packages/eland/ml/pytorch/transformers.py", line 649, in init
raise TypeError(
T**ypeError: Tokenizer type CLIPTokenizer**(name_or_path='openai/clip-vit-base-patch32', vocab_size=49408, model_max_length=77, is_fast=False, padding_side='right', truncation_side='right', special_tokens={'bos_token': AddedToken("<|startoftext|>", rstrip=False, lstrip=False, single_word=False, normalized=True), 'eos_token': AddedToken("<|endoftext|>", rstrip=False, lstrip=False, single_word=False, normalized=True), 'unk_token': AddedToken("<|endoftext|>", rstrip=False, lstrip=False, single_word=False, normalized=True), 'pad_token': '<|endoftext|>'}, clean_up_tokenization_spaces=True) not supported, must be one of: <class 'transformers.models.bart.tokenization_bart.BartTokenizer'>, <class 'transformers.models.bert.tokenization_bert.BertTokenizer'>, <class 'transformers.models.bert_japanese.tokenization_bert_japanese.BertJapaneseTokenizer'>, <class 'transformers.models.deprecated.retribert.tokenization_retribert.RetriBertTokenizer'>, <class 'transformers.models.distilbert.tokenization_distilbert.DistilBertTokenizer'>, <class 'transformers.models.dpr.tokenization_dpr.DPRContextEncoderTokenizer'>, <class 'transformers.models.dpr.tokenization_dpr.DPRQuestionEncoderTokenizer'>, <class 'transformers.models.electra.tokenization_electra.ElectraTokenizer'>, <class 'transformers.models.mobilebert.tokenization_mobilebert.MobileBertTokenizer'>, <class 'transformers.models.mpnet.tokenization_mpnet.MPNetTokenizer'>, <class 'transformers.models.roberta.tokenization_roberta.RobertaTokenizer'>, <class 'transformers.models.squeezebert.tokenization_squeezebert.SqueezeBertTokenizer'>, <class 'transformers.models.xlm_roberta.tokenization_xlm_roberta.XLMRobertaTokenizer'>
The text was updated successfully, but these errors were encountered: