Replies: 8 comments 26 replies
-
Try updating the web UI, you are at an older commit. For me it worked after installing einops:
I could only get to 7000 input tokens with 24gb vram, and the generation was super slow:
Using triton + bf16 it gets a bit faster:
These changes were necessary to + config = transformers.AutoConfig.from_pretrained(
+ Path(f"{shared.args.model_dir}/{model_name}"),
+ trust_remote_code=True
+ )
+ config.attn_config['attn_impl'] = 'triton'
+
# Load the model in simple 16-bit mode by default
if not any([shared.args.cpu, shared.args.load_in_8bit, shared.args.wbits, shared.args.auto_devices, shared.args.disk, shared.args.gpu_memory is not None, shared.args.cpu_memory is not None, shared.args.deepspeed, shared.args.flexgen, shared.model_type in ['rwkv', 'llamacpp']]):
- model = LoaderClass.from_pretrained(Path(f"{shared.args.model_dir}/{model_name}"), low_cpu_mem_usage=True, torch_dtype=torch.bfloat16 if shared.args.bf16 else torch.float16, trust_remote_code=trust_remote_code)
+ model = LoaderClass.from_pretrained(Path(f"{shared.args.model_dir}/{model_name}"), low_cpu_mem_usage=True, torch_dtype=torch.bfloat16 if shared.args.bf16 else torch.float16, trust_remote_code=trust_remote_code, config=config)
+ print("yes")
Then start with
|
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
How many memory to load thee model? INFO:Loading mosaicml_mpt-7b-storywriter... |
Beta Was this translation helpful? Give feedback.
-
amazing stuff! |
Beta Was this translation helpful? Give feedback.
-
I ran an update this morning and then attempted to run with the following slightly diff model (still mosaic that requires the remote flag) |
Beta Was this translation helpful? Give feedback.
-
A 4-bit version is now available: https://huggingface.co/OccamRazor/mpt-7b-storywriter-4bit-128g EDIT: ERROR: Can't determine model type from model name. Please specify it manually using --model_type argument The model_type in the config.json file is listed as "mpt". Is that something that would need to be added to the base webui toolkit? |
Beta Was this translation helpful? Give feedback.
-
I tried to run the
Dose it need to be changed anywhere else? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Has anyone gotten this to load into the UI?
There's a new 65K token storywriter that was just released...
https://huggingface.co/mosaicml/mpt-7b-storywriter
I added --trust-remote-code to webui.py
But still get an error:
ValueError: Loading models\mosaicml_mpt-7b-storywriter requires you to execute the configuration file in that repo on your local machine. Make sure you have read the code there to avoid malicious use, then set the option trust_remote_code=True to remove this error.
Beta Was this translation helpful? Give feedback.
All reactions