Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE] Allow local LLM backbones #15

Closed
JulianGerhard21 opened this issue Apr 20, 2023 · 4 comments
Closed

[FEATURE] Allow local LLM backbones #15

JulianGerhard21 opened this issue Apr 20, 2023 · 4 comments
Labels
type/feature Feature request

Comments

@JulianGerhard21
Copy link

馃殌 Feature

Currently there is a fixed list of LLM backbones hosted and accessible via Huggingface. It would be nice to also be able to specify a local path to a pre-trained model in the UI dropdown. Since "only" CausalLanguageModeling is currently supported, loading or training the pre-trained model is of course subject to certain conditions that would have to be checked accordingly.

Motivation

The scenarios or motivations for the feature are many. On the one hand, I create models myself with the tool, which I may want to fine-tune further. On the other hand, there are models that I try out locally and want to work with. Also, this would be a transitional solution for models located in private-flagged huggingface repositories.

What do you think?

@JulianGerhard21 JulianGerhard21 added the type/feature Feature request label Apr 20, 2023
@maxjeblick
Copy link
Contributor

Hi Julian, thanks for your interest in h2o-llmstudio!

It is possible to specify a local folder in the LLM Backbone text field (below I use the local cache folder of facebook--opt-125m, any folder containing weights/configs should also work).
We will improve the documentation and add a note about the use of local model weights.

Regarding your comment about private-flagged huggingface repositories: It should be possible to have access to private repositories once Huggingface hub access has been configured locally.

Screenshot 2023-04-20 at 10-46-58 H2O LLM Studio

@psinger
Copy link
Collaborator

psinger commented Apr 20, 2023

For continuing to train from a previous experiment, there are two additional ways to the one elaborated by @maxjeblick:

A: Use the GUI and New Experiment functionality

1.) You can re-use the settings from an old experiment by clicking on New experiment.

image

2.) You can then tick Use previous experiment weights to directly use the old weights.

image

B: Use the pretrained_weights config setting in CLI

You can just set a path to a checkpoint.pth file in pretrained_weights config setting.
This needs to be in format of how H2O LLM Studio outputs the checkpoints, bun can reference any old checkpoint you have.
We will also enable this in GUI based on this feedback as another way.

@JulianGerhard21
Copy link
Author

Awesome - and sorry for not evaluating that properly. I am working for the first time with your tool - exactly what I was looking for besides the fact that I cannot use one of the provided backbones. Thanks for answering!

@psinger
Copy link
Collaborator

psinger commented Apr 20, 2023

Thanks @JulianGerhard21 - please continue asking any questions if something is unclear.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type/feature Feature request
Projects
None yet
Development

No branches or pull requests

3 participants