Table of Best Results #2755
SilvaRaulEnrique
started this conversation in
Ideas
Replies: 2 comments 4 replies
-
Running on an M2, built from scratch, I get this error: Anyone know why this is happening? |
Beta Was this translation helpful? Give feedback.
4 replies
-
"-gqa used to be needed when loading Llama2 70B model, but in the current code with the new GGUF format it is no longer used" |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The idea of this post is construct and maintain colaborativaly a complete and actualizated best orders ranking list table to compile and executate, considering best models results for the more very long context with efectively complety coherent response, by example:
For NVidia 3090 (with CUDA and 24 GB RAM), for now the best orders are:
or?
./train-text-from-scratch
--vocab-model ServidorIA/models/s3nh/longchat-7b-v1.5-32k.ggmlv3.q8_0.bin
--ctx 64
--embd 256
--head 8
--layer 16
--checkpoint-out chk-Ultimas_Acordadas_y_Circulares-256x16.bin
--model-out ggml-Ultimas_Acordadas_y_Circulares-256x16-f32.bin
--train-data "Ultimas_Acordadas_y_Circulares.txt"
-t 4
-b 8
-n 32
--seed 1
--adam-iter 16
--print-details-interval 0
--predict 16
--use-flash
--mem-compute 8
Note: models TheBloke, and s3nh are in https://huggingface.co/
¿There is the best parameters or considerer "now" one best?, ¿and for another hardware configuration?
Beta Was this translation helpful? Give feedback.
All reactions