Description
1 update config.yml with my current used model_name and base_URL(all are located in local cloud)
2 run command according to readme.md
inference-perf --config_file config.yml --log-level DEBUG
OSError: We couldn't connect to 'https://huggingface.co' to load the files, and couldn't find them in the cached files.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/opt/conda/envs/py312/bin/inference-perf", line 8, in
sys.exit(main_cli())
^^^^^^^^^^
File "/opt/conda/envs/py312/lib/python3.12/site-packages/inference_perf/main.py", line 122, in main_cli
raise Exception("Tokenizer initialization failed") from e
not sure why it connects "https://huggingface.co" ?