You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
On my Mac, Alpaca is stuck and does not reply. Llama does respond, but Alpaca does not (see screenshot: The first request was used Llama, the second used Alpaca)
How can I cleanly uninstall everything and try installing again? I deleted the "alpaca" folder inside the "dalai" folder; and re-installed alpaca (terminal command npx dalai alpaca install 7B). I am not sure this was the correct approach.
Specs: Mac M@ 16GB Ram, 1TB storage (>800GB left space)
See below what I had on the prompt after my second install :
mkdir /Users/MYUSER/dalai
{ method: 'install', callparams: [ '7B' ] }
2 [Error: ENOENT: no such file or directory, rename '/Users/MYUSER/dalai/alpaca/models' -> '/Users/MYUSER/dalai/tmp/models'] {
errno: -2,
code: 'ENOENT',
syscall: 'rename',
path: '/Users/MYUSER/dalai/alpaca/models',
dest: '/Users/MYUSER/dalai/tmp/models'
}
3 [Error: ENOENT: no such file or directory, lstat '/Users/MYUSER/dalai/alpaca'] {
errno: -2,
code: 'ENOENT',
syscall: 'lstat',
path: '/Users/MYUSER/dalai/alpaca'
}
mkdir /Users/MYUSER/dalai/alpaca
try fetching /Users/MYUSER/dalai/alpaca https://github.com/ItsPi3141/alpaca.cpp
[E] Pull TypeError: Cannot read properties of null (reading 'split')
at new GitConfig (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/isomorphic-git/index.cjs:1610:30)
at GitConfig.from (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/isomorphic-git/index.cjs:1633:12)
at GitConfigManager.get (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/isomorphic-git/index.cjs:1756:22)
at async _getConfig (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/isomorphic-git/index.cjs:5467:18)
at async normalizeAuthorObject (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/isomorphic-git/index.cjs:5477:19)
at async Object.pull (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/isomorphic-git/index.cjs:11761:20)
at async Dalai.add (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/dalai/index.js:394:7)
at async Dalai.install (/Users/MYUSER/.npm/_npx/3c737cbb02d79cc9/node_modules/dalai/index.js:346:5) {
caller: 'git.pull'
}
try cloning /Users/MYUSER/dalai/alpaca https://github.com/ItsPi3141/alpaca.cpp
next alpaca [AsyncFunction: make]
exec: make in /Users/MYUSER/dalai/alpaca
make
exit
The default interactive shell is now zsh.
To update your account to use zsh, please run `chsh -s /bin/zsh`.
For more details, please visit https://support.apple.com/kb/HT208050.
bash-3.2$ make
I llama.cpp build info:
I UNAME_S: Darwin
I UNAME_P: arm
I UNAME_M: arm64
I CFLAGS: -I. -O3 -DNDEBUG -std=c11 -fPIC -pthread -DGGML_USE_ACCELERATE
I CXXFLAGS: -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread
I LDFLAGS: -framework Accelerate
I CC: Apple clang version 14.0.3 (clang-1403.0.22.14.1)
I CXX: Apple clang version 14.0.3 (clang-1403.0.22.14.1)
cc -I. -O3 -DNDEBUG -std=c11 -fPIC -pthread -DGGML_USE_ACCELERATE -c ggml.c -o ggml.o
c++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread -c utils.cpp -o utils.o
c++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread main.cpp ggml.o utils.o -o main -framework Accelerate
./main -h
usage: ./main [options]
options:
-h, --help show this help message and exit
-i, --interactive run in interactive mode
--interactive-start run in interactive mode and poll user input at startup
-r PROMPT, --reverse-prompt PROMPT
in interactive mode, poll user input upon seeing PROMPT
--color colorise output to distinguish prompt and user input from generations
-s SEED, --seed SEED RNG seed (default: -1)
-t N, --threads N number of threads to use during computation (default: 4)
-p PROMPT, --prompt PROMPT
prompt to start generation with (default: random)
-f FNAME, --file FNAME
prompt file to start generation.
-n N, --n_predict N number of tokens to predict (default: 128)
--top_k N top-k sampling (default: 40)
--top_p N top-p sampling (default: 0.9)
--repeat_last_n N last n tokens to consider for penalize (default: 64)
--repeat_penalty N penalize repeat sequence of tokens (default: 1.3)
-c N, --ctx_size N size of the prompt context (default: 2048)
--temp N temperature (default: 0.1)
-b N, --batch_size N batch size for prompt processing (default: 8)
-m FNAME, --model FNAME
model path (default: ggml-alpaca-7b-q4.bin)
c++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread quantize.cpp ggml.o utils.o -o quantize -framework Accelerate
bash-3.2$ exit
exit
alpaca.add [ '7B' ]
dir /Users/MYUSER/dalai/alpaca/models/7B
downloading torrent
ggml-model-q4_0.bin 100%[==================================================================================>] done
The text was updated successfully, but these errors were encountered:
On my Mac, Alpaca is stuck and does not reply. Llama does respond, but Alpaca does not (see screenshot: The first request was used Llama, the second used Alpaca)
How can I cleanly uninstall everything and try installing again? I deleted the "alpaca" folder inside the "dalai" folder; and re-installed alpaca (terminal command npx dalai alpaca install 7B). I am not sure this was the correct approach.
Specs: Mac M@ 16GB Ram, 1TB storage (>800GB left space)
See below what I had on the prompt after my second install :
The text was updated successfully, but these errors were encountered: