-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How do I add a model to run it locally? #851
Comments
Did you click the top button? i.e. "Download/Load Model" We are aware the coloring is not great. It's a limitation of gradio that buttons and info labels are same. We have some work on it: #818 |
Hi @pseudotensor , jap I did that, but still same error as above |
I presume it's not finding the file. I haven't had issues. Are you able to diagnose? |
I am having a similar issue but with the installation from the One Click Windows GPU CUDA Installer. I can't figure out how to load a model. Do I need to place the .bin file at a specific location? When I select something from the dropdown and click Load-Unload Model / LORA button on the right, I get this error on the top right: |
@natlamir It seems to have trouble writing some files, probably permissions issue to the disk for where it was installed. One can debug like this: #652 (comment) i.e. using python instead of pythonx and running on windows command line termninal. If you use llama as base_model, then you can provide GGML link from TheBloke. I give details here: https://github.com/h2oai/h2ogpt/blob/main/docs/FAQ.md#adding-models But your issue is some permission thing, a stack trace from the command line output would help. Thanks! |
@pseudotensor Thanks for the fast reply. I tried running it through the command line to get the stack trace, and it works just fine when run through the command line! (I was using a non-elevated command prompt) Previously I was trying to run it by clicking on the icon from the Start menu on my Windows 10, and that is when it was erroring. So now I am able to download and use the model I was trying to. I tried using the GGML link from TheBloke you mentioned. Let me know if I am missing a step for doing this through the UI. In the Models tab, on the bottom left textbox titled "New Model name/path/URL" I enter: https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/resolve/main/llama-2-7b-chat.ggmlv3.q8_0.bin Then click on the "Add new Model, Lora, Server url:port" button on the bottom right. This auto-populated the URL in the "Choose Base Model" dropdown at the top with the URL I entered in the textbox. Then I click the "Load-Unload Model/LORA" button on the top right, and it downloads the 7GB file, but then errors. Here is the command line output / stack trace of the error (The file it is referencing in the error in the Temp folder appears to be the model file without an extension. it is 7GB): C:\Users\root>C:\Users\root\AppData\Local\Programs\h2oGPT\Python\python.exe "C:\Users\root\AppData\Local\Programs\h2oGPT\h2oGPT.launch.pyw" To create a public link, set During handling of the above exception, another exception occurred: Traceback (most recent call last): |
You can't pass GGML model to --base_model. See: https://github.com/h2oai/h2ogpt/blob/main/docs/FAQ.md#adding-models For GGML, use 'llama' as base_model, then in UI you'll get more options appear. Then put in the url for the model llama name. |
For general offline, see updates here: https://github.com/h2oai/h2ogpt/blob/main/docs/README_offline.md from #877 |
The "base_model" is llama for that model. Once you choose "llama" another view will pop-up to enter the llama model path or url. Then you click on the "Download/Load Models" button at top. We'll try to improve the UX. |
Hi, i am using a ubuntu system in laptop how many times i have tried to install h2o.ai in local through terminal i have been facing lot of issues and i could't get what the mistake happens with system can help me |
@Blue-newai if you have a problem you should post the issue etc. I've tried to make it easier and easier to install and use. |
Dear community,
![h20gpt](https://private-user-images.githubusercontent.com/9105222/268204607-db34ab05-04e7-436f-9a4d-8d82618e8b10.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTk3Njc0MTksIm5iZiI6MTcxOTc2NzExOSwicGF0aCI6Ii85MTA1MjIyLzI2ODIwNDYwNy1kYjM0YWIwNS0wNGU3LTQzNmYtOWE0ZC04ZDgyNjE4ZThiMTAucG5nP1gtQW16LUFsZ29yaXRobT1BV1M0LUhNQUMtU0hBMjU2JlgtQW16LUNyZWRlbnRpYWw9QUtJQVZDT0RZTFNBNTNQUUs0WkElMkYyMDI0MDYzMCUyRnVzLWVhc3QtMSUyRnMzJTJGYXdzNF9yZXF1ZXN0JlgtQW16LURhdGU9MjAyNDA2MzBUMTcwNTE5WiZYLUFtei1FeHBpcmVzPTMwMCZYLUFtei1TaWduYXR1cmU9Y2YwNmRlMjQwMTEyMWQxMTY5NTExMTJmMzMwYWVjM2UzYWY3MjY5Y2M0MjQ1NWUwZjc1NTQ2ZDMwY2Q2OTQ3OSZYLUFtei1TaWduZWRIZWFkZXJzPWhvc3QmYWN0b3JfaWQ9MCZrZXlfaWQ9MCZyZXBvX2lkPTAifQ.Metbg3A8RhmACbagt9ZJM66GNgYjs3ssPmIOO33MlUk)
I have set the repo up and running on my windows machine. However I do not udnerstand how I can add a model to run it locally on my machine. I have downloaded the model 'llama-2-7b-chat.ggmlv3.q8_0.bin' and placed it into the h20GPT root folder. Then in the UI I have selcted it like so:
But if I entered anything in the prompt I get the error on the console:
AssertionError: Please choose a base model with --base_model (CLI) or load in Models Tab (gradio). Then start New Conversation
Can anyone point me into the right direction on how to do this correctly? Thank's a lot!
The text was updated successfully, but these errors were encountered: