Skip to content
guinmoon edited this page Jun 17, 2024 · 6 revisions

FAQ

How to install

At the moment the application is available in TestFlight. If for some reason you are not able to access testflight email me and i will build an ipa for you.

What model formats are supported

Models in ggjtv3 (must have a .bin extension) and gguf (must have a .gguf extension) formats are supported.

Model_load_error

Make sure the correct inference is selected in the chat settings. If the inference is correct, but the error persists, it is possible that the model is in an old format and needs to be converted and requatized from the original model. Scripts for inference and quantization you can find here: LLaMA, LLaMA 2, Falcon, Starcoder Alpaca, GPT4All, Chinese LLaMA, Vigogne, Vicuna, Koala, OpenBuddy, Pygmalion 7B, WizardLM, Baichuan-7B, Aquila-7B, Mistral

GPT-2, GPT-NEOX, Replit

Eval error

If the output starts normally but after a while the eval error occurs, try increasing the size of the context in chat settings.

Application crashes without warning

The most common reason for an application crash is lack of memory. You may be using a model that is too large.

Why Extended Virtual Addressing Entitlement required?

Extended Virtual Addressing and increased-memory-limit entitlements is required to run models larger than 3B, otherwise the application may crash. If for some reason you cannot build an application with this entitlement, you can simply turn it off.

extras.otherInstrOffset != 0 && "Kind::arm64_adrp_ldr missing extra info"

Solution