https://github.com/OpenLemur/Lemur
随着大量的大型语言模型(LLM)和聊天机器人每周都在发布,通常都对其性能进行了浮夸的宣称,很难筛选出开源社区正在取得的真正进展,以及哪种模型是当前的技术水平🤗 开放式LLM排行榜旨在跟踪、排名和评估LLM和聊天机器人的发布情况。我们在Eleuther AI语言模型评估线束的4个关键基准上评估模型,Eleuther人工智能语言模型评估套件是一个统一的框架,用于在大量不同的评估任务上测试生成语言模型。该排行榜的一个关键优势是,社区中的任何人都可以在🤗 GPU集群,只要它是🤗 轮毂上有重物的变压器模型。我们还支持为非商业许可模型(如LLaMa)评估具有增量权重的模型。
Model | Average ⬆️ | ARC ⬆️ | HellaSwag ⬆️ | MMLU ⬆️ | TruthfulQA (MC) ⬆️ | Type | Hub License | #Params (B) | Hub ❤️ | Model sha | model_name_for_query |
---|---|---|---|---|---|---|---|---|---|---|---|
67 | 64.9 | 84.9 | 61.9 | 56.3 | N/A | 30 | 48 | be44a37814a20e790063086703f570732597887a | upstage/llama-30b-instruct-2048 | ||
66.8 | 64.6 | 85.9 | 63.9 | 52.8 | N/A | 68.977 | 406 | 7f54101c0fbb67a8143ca23eb8bd09b71f269c74 | meta-llama/Llama-2-70b-chat-hf | ||
66.6 | 65.8 | 84.8 | 63.5 | 52.4 | N/A | other | 30 | 8 | 836cf4dcd60ebe2ff09415c72f809d94639e8d35 | lilloukas/GPlatty-30B | |
66.4 | 65.8 | 83.9 | 62.6 | 53.5 | N/A | other | 30 | 4 | 017e1c32bca060107337dbf26db2044a7caa56f2 | ariellee/SuperPlatty-30B | |
66.1 | 64.9 | 84.3 | 56.5 | 58.6 | finetuned | 30 | 103 | 24da9e88f2b2b7946bc6fe9412d6728b9adc2c3d | CalderaAI/30B-Lazarus | ||
65.4 | 61.4 | 85.2 | 69.3 | 45.7 | N/A | 68.977 | 239 | b70e3df5fc2bf95ba5859878cafa2da37908598c | meta-llama/Llama-2-70b-hf | ||
65.2 | 62.5 | 86.2 | 59.4 | 52.8 | N/A | 30 | 10 | fea4312379557e8a1e8073965f560798de369edd | upstage/llama-30b-instruct | ||
65.1 | 64.8 | 85.1 | 56.6 | 54 | finetuned | 30 | 122 | dc9d81f454d286ea040c5cd45b058aecaa51c13e | ausboss/llama-30b-supercot | ||
64.6 | 64.6 | 84.2 | 64.2 | 45.4 | N/A | other | 30 | 13 | 979ad39b58a8e4a9419b7bc7a0dc8419f3912e71 | lilloukas/Platypus-30B | |
64.4 | 64.2 | 84.4 | 57.5 | 51.6 | finetuned | 32.529 | 12 | 0cff8e9718e57202171003d556d2e6630061879d | openaccess-ai-collective/manticore-30b-chat-pyg-alpha | ||
64.3 | 64.5 | 85.2 | 59.1 | 48.4 | N/A | 30 | 17 | 64c10edf5312cd13704925b07413882d9e94c7a0 | openaccess-ai-collective/hippogriff-30b-chat | ||
64.3 | 63 | 83.7 | 58.2 | 52.3 | N/A | other | 30 | 3 | e2329c05a6e59660ba3cbcc01adf30a78f852594 | concedo/Vicuzard-30B-Uncensored | |
64 | 63.1 | 83.6 | 57.7 | 51.5 | finetuned | 30 | 59 | 1a0d1d72a40946463fb4a9780207da19bfecc38b | MetaIX/GPT4-X-Alpasta-30b | ||
64 | 61.8 | 85.1 | 57.5 | 51.5 | N/A | 30 | 2 | 2424b6346e9e8fd749b9a6734f5d7125b5926daf | dsvv-cair/alpaca-cleaned-llama-30b-bf16 | ||
63.6 | 61.3 | 83.8 | 57.9 | 51.2 | N/A | apache-2.0 | 30 | 4 | e04207847429af03c4780f5ac85c726536217981 | Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-hf | |
63.6 | 61.9 | 83.8 | 57.6 | 51 | finetuned | 30 | 7 | c63d117d1ec5794766dd6dc5e1469769df8aba1d | Aeala/VicUnlocked-alpaca-30b | ||
63.6 | 62.1 | 83.4 | 58.2 | 50.8 | N/A | other | 30 | 54 | 6374baef4cedd41f85c111b8eec3eb38ee24c4b9 | ehartford/Wizard-Vicuna-30B-Uncensored | |
63.6 | 62.1 | 83.4 | 58.2 | 50.8 | N/A | other | 30 | 11 | c7b7cecb5a314fc66deebabcb67c230a3fbe84f7 | TheBloke/Wizard-Vicuna-30B-Uncensored-fp16 | |
63.4 | 62.1 | 82.8 | 56.2 | 52.7 | N/A | 30 | 1 | ee76c821f861f0ab0276f9f429dd06565f1f2051 | Aeala/GPT4-x-AlpacaDente-30b | ||
63.4 | 61.6 | 84.3 | 55.4 | 52.5 | finetuned | apache-2.0 | 40 | 1009 | 7475ff8cfc36ed9a962b658ae3c33391566a85a5 | tiiuae/falcon-40b-instruct | |
63.4 | 61.5 | 82.1 | 54.2 | 55.9 | N/A | cc-by-nc-4.0 | 30 | 3 | aa9912a2ac60abeac28b4566731cd903dcc582ac | MayaPH/GodziLLa-30B | |
63 | 62.5 | 84.5 | 53.8 | 51.2 | finetuned | 33 | 161 | b2e78a916582935b6616d184b22ea5e9e1eb4c34 | timdettmers/guanaco-33b-merged | ||
62.9 | 60.2 | 82.9 | 56.8 | 51.6 | N/A | other | 30 | 110 | 761783745fcb97831ad8035d3cbd5de484aca3ce | ehartford/WizardLM-30B-Uncensored | |
62.6 | 62.7 | 85 | 58.5 | 44.2 | finetuned | other | 30 | 64 | 300bc5f3dc129a3d17adf059394e381eff7fbd55 | digitous/Alpacino30b | |
62.4 | 59.6 | 79.9 | 54.4 | 55.9 | N/A | 30 | 23 | e58bafedf660477c206ad64f3118a571951bb28e | Monero/WizardLM-Uncensored-SuperCOT-StoryTelling-30b | ||
62.1 | 62.2 | 83.5 | 55.9 | 46.7 | finetuned | other | 33 | 21 | 3c11f81d9180618f13777276b1eb0eb70ab99cf0 | elinas/chronos-33b | |
62.1 | 57.6 | 84.3 | 63.4 | 43 | pretrained | other | 65.286 | 0 | 4ae2e56610e8b9b9a78472708390668e9096b4f9 | huggingface/llama-65b | |
62.1 | 57.6 | 84.3 | 63.4 | 43 | pretrained | other | 65.286 | 53 | 49707c5313d34d1c5a846e29cf2a2a650c22c8ee | huggyllama/llama-65b | |
61.9 | 60.6 | 82.2 | 57.9 | 46.9 | N/A | other | 30 | 13 | a7a2306b9a63de2c545f35b24735f4540baf5903 | TheBloke/OpenAssistant-SFT-7-Llama-30B-HF | |
61.8 | 60.6 | 81.8 | 56.6 | 48.4 | finetuned | 30 | 29 | 9fe5a8dada738f44e7ee9293b2140ae0be021787 | Aeala/GPT4-x-AlpacaDente2-30b | ||
61.7 | 61.3 | 84.7 | 58.5 | 42.3 | pretrained | other | 32.529 | 0 | 13c77caa472bfa79d4f3f0ec82cbdc9dd88e5d22 | huggingface/llama-30b | |
61.7 | 61.3 | 84.7 | 58.5 | 42.3 | N/A | apache-2.0 | 30 | 1 | fba493af11a73cf5a2ee7857dd7aecb98c659dc4 | Yhyu13/llama-30B-hf-openassitant | |
61.5 | 61.9 | 85.3 | 57 | 41.7 | pretrained | apache-2.0 | 40 | 2122 | c47b371b31a68349c233104050ac76680b8485db | tiiuae/falcon-40b | |
61.2 | 58.7 | 82.5 | 51.2 | 52.4 | N/A | cc-by-nc-sa-4.0 | 30 | 157 | 54f33278a04aa4e612bca482b82f801ab658e890 | mosaicml/mpt-30b-chat | |
61.1 | 58.3 | 82.4 | 51.2 | 52.6 | N/A | cc-by-nc-sa-4.0 | 30 | 157 | 7debc3fc2c5f330a33838bb007c24517b73347b8 | mosaicml/mpt-30b-chat | |
60.9 | 63.9 | 85.7 | 58.3 | 35.7 | N/A | apache-2.0 | 30 | 29 | 5818a6344f48dc5a324589b57cb288a9d54c0b79 | ehartford/based-30b | |
60.4 | 58.1 | 80.8 | 63.3 | 39.3 | finetuned | other | 65 | 19 | 3fa4546259d6bbd6b5d637484c325ab19181a73c | TheBloke/dromedary-65b-lora-HF | |
60.4 | 57.3 | 80.9 | 52.9 | 50.5 | N/A | other | 13 | 6 | b79733805e98e668ff9a459975c259881b1b8014 | TheBloke/wizardLM-13B-1.0-fp16 | |
60.3 | 57.6 | 80.2 | 51.9 | 51.6 | N/A | other | 13 | 4 | 734f5641f6c548474517d1536c46024517f120e0 | TheBloke/UltraLM-13B-fp16 | |
60.3 | 55.5 | 80.4 | 54 | 51.3 | N/A | other | 30 | 15 | f25d922536e602af035d476a287b68361fdda5de | Monero/WizardLM-30B-Uncensored-Guanaco-SuperCOT-30b | |
60.2 | 58.3 | 81 | 50 | 51.6 | N/A | other | 13 | 11 | 9219b61a0e8bc880e4cd0f8bebc48a97ee0950c7 | TheBloke/airoboros-13B-HF | |
60.2 | 56.6 | 82.1 | 50.4 | 51.5 | finetuned | gpl | 13 | 344 | 24e8c03148ffd1f3e469744dfc24ad2ad82848f8 | NousResearch/Nous-Hermes-13b | |
60.2 | 58.3 | 81 | 50 | 51.6 | finetuned | cc-by-nc-4.0 | 13 | 90 | 44830f9e1559f318f5dad875bab40d1d1beddbfc | jondurbin/airoboros-13b | |
60.1 | 59 | 81.9 | 47.9 | 51.7 | finetuned | other | 13 | 174 | fff9ac7f0e2e7b340f2301f5f089d989fc03be67 | TheBloke/Wizard-Vicuna-13B-Uncensored-HF | |
60.1 | 58.7 | 81.6 | 50.8 | 49.2 | finetuned | 13 | 102 | aed786b0200251c9962ac200c50f7e367f264b46 | openaccess-ai-collective/manticore-13b | ||
60.1 | 59 | 81.9 | 47.9 | 51.7 | N/A | other | 13 | 192 | 95bfd1640a54e76b3e857c2462fd3a77eca0b275 | ehartford/Wizard-Vicuna-13B-Uncensored | |
60 | 54.6 | 80.4 | 52.9 | 52.1 | N/A | 13 | 126 | 7900eeb715a49affee9e6390f824e62eea3f3fb1 | lmsys/vicuna-13b-v1.3 | ||
59.9 | 59 | 81.9 | 54.6 | 44.1 | N/A | 13.016 | 171 | f848cf15ab9a51ae5735ab28120a9a0773eeb541 | meta-llama/Llama-2-13b-chat-hf | ||
59.8 | 57.3 | 81.1 | 50.6 | 50.2 | finetuned | 13 | 99 | 76e90314541be6cfa2b55208831c99f1351c1a33 | openaccess-ai-collective/wizard-mega-13b | ||
59.8 | 56.8 | 82.3 | 47.8 | 52.3 | N/A | 13 | 13 | de665d6002f1f2ef460408e8fa5bc1e0b7bb99b1 | Monero/Manticore-13b-Chat-Pyg-Guanaco | ||
59.8 | 59.9 | 81.1 | 47.2 | 51.2 | N/A | 13 | 7 | 90c02cc338afcdd890a948af06432674743363ad | TehVenom/Metharme-13b-Merged | ||
59.8 | 59.7 | 83.3 | 47.5 | 48.8 | N/A | cc-by-nc-4.0 | 13 | 14 | d0d2687ed2b4a63a644ed6c5b3f6401844718659 | jondurbin/airoboros-13b-gpt4-1.4 | |
59.7 | 58.4 | 81.6 | 49.4 | 49.5 | N/A | other | 13 | 8 | d89d925ad1eeaee465c4de3e5c74240a5a40b585 | Gryphe/MythoLogic-13b | |
59.7 | 58.5 | 82.2 | 50.6 | 47.5 | N/A | other | 13 | 25 | aa828ef92c363a5577ffd7d29e678277b9d2eb3c | digitous/13B-HyperMantis | |
59.6 | 53.4 | 80.1 | 51.2 | 53.6 | N/A | other | 13 | 20 | a247577c882940e0c6b040fe8239d760c0d10d40 | TheBloke/gpt4-x-vicuna-13B-HF | |
59.6 | 59.6 | 82.1 | 47.5 | 49 | N/A | other | 13 | 3 | 49678a2dd15fb4e1f1b99616ccc1ffd269912833 | TheBloke/gpt4-alpaca-lora-13B-HF | |
59.5 | 58.5 | 82 | 48.8 | 48.8 | N/A | 13.016 | 20 | f9ef65a3cf50e3c09ccb443f99225148e08517aa | openaccess-ai-collective/manticore-13b-chat-pyg | ||
59.4 | 58.5 | 79.9 | 46 | 53.1 | N/A | 13 | 4 | 50af05b015446110a2dc52a1b4b341142c98e62b | Aeala/GPT4-x-Alpasta-13b | ||
59.4 | 55.1 | 79.7 | 50.1 | 52.6 | N/A | cc-by-nc-sa-4.0 | 13 | 30 | 1058709314f7ca090937d0a2b7b37b0b3a8f12a3 | psmathur/orca_mini_v2_13b | |
59.4 | 59 | 79.7 | 49.1 | 49.6 | N/A | 13 | 0 | dd326f89ce885844d714d9ab33603e0d17f56cc5 | YeungNLP/firefly-llama-13b | ||
59.2 | 52.7 | 80.1 | 51.9 | 52.1 | finetuned | 13 | 94 | 8c71dbe9221e83d2ec72e4dc08beccfc78b563c0 | TheBloke/vicuna-13B-1.1-HF | ||
59.2 | 52.7 | 80.1 | 51.9 | 52.1 | N/A | 13 | 3 | 346e3c46959cf9f1e03feffa761afe020c0fb6a8 | kevinpro/Vicuna-13B-CoT | ||
59.2 | 52.7 | 80.1 | 51.9 | 52.1 | N/A | apache-2.0 | 13 | 132 | bfcc6ca66694310be6c85ba0638597f4256c4143 | eachadea/vicuna-13b-1.1 | |
59.2 | 52.7 | 80.1 | 51.9 | 52.1 | finetuned | 13 | 0 | 83fa0860990df1db35550f973ba4306449e35412 | pillowtalks-ai/delta13b | ||
59.1 | 57.8 | 83.8 | 48.3 | 46.7 | N/A | other | 13 | 6 | bd59c700815124df616a17f5b49a0bc51590b231 | TheBloke/guanaco-13B-HF | |
59 | 57.9 | 82.4 | 48.6 | 47.3 | N/A | 13 | 8 | 0e3796192f7edf43968541b9454ea35da4a2b1c5 | dvruette/oasst-llama-13b-2-epochs | ||
58.8 | 51.7 | 79.9 | 50.8 | 52.7 | finetuned | 13 | 96 | ac4218770a58baaaaf25201076fe082abb6ffd13 | eachadea/vicuna-13b | ||
58.7 | 59.4 | 82.1 | 55.8 | 37.4 | N/A | 13.016 | 125 | 7da18fb10421c3ae2a1eb92815bad75e84816e35 | meta-llama/Llama-2-13b-hf | ||
58.7 | 56.4 | 79.1 | 49.6 | 49.6 | finetuned | apache-2.0 | 13 | 8 | b5ae4519d4c8f4559a0aa80b6efe2008413ece01 | openaccess-ai-collective/minotaur-13b | |
58.6 | 53.8 | 79.5 | 51.5 | 49.5 | N/A | other | 13 | 0 | 388bc2f82a1ee8b963c7f94f9c7b6743f7214306 | wahaha1987/llama_13b_sharegpt94k_fastchat | |
58.5 | 59 | 80.3 | 47.2 | 47.4 | N/A | 13 | 6 | 6cb016f5bfcbc24ee08312b52f08ef5e8f860871 | dvruette/llama-13b-pretrained-sft-do2 | ||
58.4 | 58.4 | 81.7 | 47.9 | 45.4 | finetuned | 13 | 4 | 3a82b04684fe99d59556421c3f96a187049a3cec | xzuyn/Alpacino-SuperCOT-13B | ||
58.4 | 56.9 | 79.3 | 49.7 | 47.9 | finetuned | cc-by-nc-4.0 | 13 | 19 | a3c4bbccca8b650700a49a225582c17bb49b446b | project-baize/baize-v2-13b | |
58.1 | 54.9 | 78.2 | 50.1 | 49.1 | N/A | mit | 13 | 125 | 4c558283a98348383460939afda9cb5c54544c8f | Open-Orca/OpenOrca-Preview1-13B | |
58.1 | 56.4 | 79.3 | 48.4 | 48.4 | N/A | 13 | 0 | cc5ee2231066c147423f89e9df40f7364c3275a5 | haonan-li/bactrian-x-llama-13b-merged | ||
58.1 | 54.7 | 79.2 | 48.9 | 49.6 | N/A | 13 | 48 | 12dc8aacb474522ae2a83c18cb0fdf0907987f8f | TheBloke/wizard-vicuna-13B-HF | ||
58.1 | 54.7 | 79.2 | 48.9 | 49.6 | finetuned | 13 | 68 | 419dc5acc391de54a60d0b041e94e767d1ef2032 | junelee/wizard-vicuna-13b | ||
58 | 55.6 | 79.2 | 49.7 | 47.4 | N/A | 13 | 11 | 6d98f2801f13d89de7978ee9f348a52ea46a24ec | camel-ai/CAMEL-13B-Combined-Data | ||
57.9 | 53.9 | 80.7 | 53.2 | 43.8 | N/A | other | 13 | 2 | 532aeb363b0ceee155b3cf9479ef635b797cee7c | TheBloke/tulu-13B-fp16 | |
57.9 | 56.1 | 81.7 | 45.4 | 48.5 | finetuned | other | 13 | 8 | f6953fa162b487a3d4c6bdc7b7951e09576c2ae5 | ausboss/llama-13b-supercot | |
57.8 | 56.3 | 79.3 | 47 | 48.4 | N/A | 13 | 0 | c28cc0cf5a1a1bf4de96b23d06b02129dca85eb9 | dvruette/llama-13b-pretrained | ||
57.7 | 56.4 | 79.3 | 46.6 | 48.6 | N/A | 13 | 1 | 045c84727d495bfb4b612a2482ce0d807c067b46 | dvruette/llama-13b-pretrained-dropout | ||
57.5 | 58.4 | 84.3 | 49.2 | 38 | N/A | cc-by-sa-3.0 | 30 | 86 | 2abf1163dd8c9b11f07d805c06e6ec90a1f2037e | mosaicml/mpt-30b-instruct | |
57.5 | 58.8 | 81.8 | 48.1 | 41.2 | finetuned | 13 | 0 | a7e5484df8aceae7800ae9301a3954cf74b527e9 | PocketDoc/Dans-PileOfSets-Mk1-llama-13b-merged | ||
57.4 | 58.5 | 81.3 | 47.9 | 41.7 | N/A | other | 13 | 27 | 7092a5c8dec649694dd66ff8cfe5452ce52e6a40 | digitous/Alpacino13b | |
57.4 | 58.4 | 84.1 | 48.9 | 38.1 | N/A | cc-by-sa-3.0 | 30 | 86 | 2abf1163dd8c9b11f07d805c06e6ec90a1f2037e | mosaicml/mpt-30b-instruct | |
56.8 | 57.3 | 80 | 45.5 | 44.5 | N/A | 13 | 0 | 1f839c019153789c15bbc45ecbb512d0f5015881 | dvruette/llama-13b-pretrained-sft-epoch-1 | ||
56.8 | 54.9 | 79.2 | 46.6 | 46.3 | N/A | 13 | 10 | 762ecb0d85572c8f8bcbca06d27f7f64a4d74615 | camel-ai/CAMEL-13B-Role-Playing-Data | ||
56.5 | 53 | 77.6 | 45.3 | 50.2 | N/A | other | 13 | 38 | b20f96a0171ce4c0fa27d6048215ebe710521587 | TheBloke/koala-13B-HF | |
56.4 | 53.7 | 77.5 | 45.6 | 48.9 | finetuned | apache-2.0 | 7 | 103 | 9d8eea215e00b388a22e8f050768ea8911d41f1d | eachadea/vicuna-7b-1.1 | |
56.4 | 53.7 | 77.5 | 45.6 | 48.9 | N/A | 7 | 1 | d971d788db19648ad16bf77ec3f1de35ebf9a8e0 | Ejafa/vicuna_7B_vanilla_1.1 | ||
56.4 | 52.9 | 78.6 | 48.3 | 45.6 | N/A | 6.738 | 225 | b7701a9e825e79a5ab18b5801be113c2160cc627 | meta-llama/Llama-2-7b-chat-hf | ||
56.2 | 55.9 | 82.4 | 47.9 | 38.4 | N/A | apache-2.0 | 30 | 291 | 0261af71d7177453889f868d26607dec8d5aaa2e | mosaicml/mpt-30b | |
56.1 | 56.2 | 80.9 | 47.7 | 39.5 | pretrained | other | 13.016 | 0 | 4022c52fcc7473ce7364bb5ac166195903ea1efb | huggingface/llama-13b | |
56 | 58.1 | 81.5 | 48.6 | 36 | N/A | 13 | 0 | d2cd599cc40db3370009f45d6caa7e486cb6d31f | dvruette/oasst-llama-13b-1000-steps | ||
55.8 | 52.5 | 80.6 | 42.8 | 47.2 | finetuned | bigcode-openrail-m | 15.517 | 156 | b1bcda690655777373f57ea6614eb095ec2c886f | HuggingFaceH4/starchat-beta | |
55.7 | 54.9 | 82.3 | 47.4 | 38.2 | N/A | apache-2.0 | 30 | 291 | f54ab0cc237fafbc9afc10d285e0fa85299c9bd8 | mosaicml/mpt-30b | |
55.6 | 50.4 | 76.9 | 48.1 | 47 | N/A | 7 | 82 | ac066c83424c4a7221aa10c0ebe074b24d3bcdb6 | lmsys/vicuna-7b-v1.3 | ||
55.3 | 56.1 | 80.3 | 47.9 | 37 | N/A | agpl-3.0 | 13 | 3 | 142e198df473fd0cd4370b0d50be5f57e1da399b | llama-anon/instruct-13b | |
55.2 | 53.2 | 73.5 | 48.8 | 45.3 | N/A | other | 13 | 28 | a118d2c35573b9a70c6f5b56fba4b657f74ce00c | shibing624/chinese-alpaca-plus-13b-hf | |
55 | 53.2 | 76.9 | 44.6 | 45.3 | finetuned | other | 7 | 1 | 2d82abff150b7a5ae484f9cd7c64c72fd4eaf7f5 | wahaha1987/llama_7b_sharegpt94k_fastchat | |
54.6 | 50.9 | 76.6 | 44 | 46.7 | N/A | other | 13 | 406 | 9025c5f96fef9525da9238369ad082961b0e9494 | ehartford/WizardLM-13B-Uncensored | |
54.4 | 53.5 | 78.4 | 44.6 | 41.3 | finetuned | 7 | 10 | 0b3ef975fb5e8ac1eae775160ab54c98221889df | xzuyn/MedicWizard-7B | ||
54.4 | 53.1 | 78.6 | 46.9 | 38.8 | N/A | 6.738 | 150 | e8f058fa738b6b308540024e9aa12e274e291f75 | meta-llama/Llama-2-7b-hf | ||
54.1 | 50.2 | 77 | 47.6 | 41.6 | N/A | other | 7 | 1 | 8a026683f79119643f4007da4e9155c7849792cc | TheBloke/tulu-7B-fp16 | |
54.1 | 54.1 | 80.4 | 41.5 | 40.5 | finetuned | cc | 7 | 22 | b57b9f5ff34059e485b769973d023021fc66a8f7 | medalpaca/medalpaca-7b | |
53.8 | 56.5 | 80 | 42.9 | 35.9 | N/A | 13 | 21 | f96308083033c84db47b6c093da3817c085c87c7 | TehVenom/Pygmalion-13b-Merged | ||
53.2 | 53.4 | 78.8 | 37.1 | 43.5 | N/A | other | 7 | 48 | 1097285acd9c48a1d09bc0a9844d365384732111 | ehartford/Wizard-Vicuna-7B-Uncensored | |
53.2 | 55 | 78.8 | 37.5 | 41.5 | finetuned | 7 | 5 | bbece5e3f8ee9be09c8defc536a95c6ef780c681 | LLMs/AlpacaGPT4-7B-elina | ||
53.2 | 53.4 | 78.8 | 37.1 | 43.5 | N/A | other | 7 | 16 | b802f1b4401d0b2242137160c20cc11b9ffd3a4c | TheBloke/Wizard-Vicuna-7B-Uncensored-HF | |
52.9 | 53.1 | 78.7 | 38.9 | 40.7 | N/A | other | 7 | 3 | 14aa50fba9f6418c0d5e2d24087eb802931040ef | TheBloke/airoboros-7b-gpt4-fp16 | |
52.8 | 53 | 77.7 | 37.3 | 43.4 | finetuned | cc-by-nc-4.0 | 7 | 13 | 7ea67f85ff3a7a8ec77f1819dec3e56779b764b1 | jondurbin/airoboros-7b | |
52.6 | 50.8 | 76 | 39.5 | 43.9 | N/A | cc-by-nc-sa-4.0 | 7 | 30 | 165850882991d7fa4eabab577a03ed84e0713bfa | psmathur/orca_mini_v2_7b | |
52.5 | 49.8 | 75.5 | 39.1 | 45.7 | N/A | cc-by-nc-sa-4.0 | 7 | 4 | 175965f50907c6a8cd40f1a4b10d28342969c066 | jphme/orca_mini_v2_ger_7b | |
52.4 | 51.5 | 76.9 | 33.3 | 48.1 | N/A | other | 6.738 | 14 | c03ac527360663d17bb142405251028eec843ed9 | notstoic/PygmalionCoT-7b | |
52.3 | 50.3 | 75.3 | 38.1 | 45.6 | finetuned | other | 7 | 83 | a8e22531a48cece989e670f539eb18ebd2dbd0cf | TheBloke/wizardLM-7B-HF | |
51.9 | 53 | 80.1 | 35.3 | 39.2 | N/A | other | 7 | 6 | 293c24105fa15afa127a2ec3905fdc2a0a3a6dac | TheBloke/guanaco-7B-HF | |
51.9 | 48.7 | 77.3 | 43.7 | 37.9 | finetuned | 142 | 95be82087c33f14ee9941c812a154a9dd66efe72 | bigcode/starcoderplus | |||
51.9 | 53.7 | 78.6 | 35.9 | 39.2 | finetuned | 6.738 | 12 | 62ca156891feead8db117be8f5f35687b6274e6e | Neko-Institute-of-Science/metharme-7b | ||
51.4 | 53.2 | 76.9 | 35.9 | 39.4 | N/A | 7 | 2 | 1f61442e1238062095b31b4909c5e9ab26105794 | TehVenom/Pygmalion_AlpacaLora-7b | ||
51.3 | 49 | 75.1 | 39.6 | 41.4 | finetuned | cc-by-nc-4.0 | 7 | 18 | e4731c2c2671e2d0b47b5eba08c753ca21671fab | project-baize/baize-v2-7b | |
51.3 | 51.3 | 77.5 | 33.2 | 43.3 | N/A | 1 | c1068f859a225e50d9d9ec74c572bfaf38573051 | wannaphong/openthaigpt-0.1.0-beta-full-model_for_open_llm_leaderboard | |||
50.7 | 49.7 | 71.1 | 43.8 | 38.3 | N/A | apache-2.0 | 13 | 418 | b6d7fde8392250730d24cc2fcfa3b7e5f9a03ce8 | openlm-research/open_llama_13b | |
50 | 52.3 | 77.5 | 36.6 | 33.8 | N/A | 0 | d985610bef080473e40f01c53266083c5f0c3169 | illuin/test-custom-llama | |||
50 | 46.5 | 75.5 | 37.6 | 40.2 | N/A | cc-by-nc-sa-4.0 | 7 | 470 | 64e5c9c9fb53a8e89690c2dee75a5add37f7113e | mosaicml/mpt-7b-chat | |
49.8 | 51.4 | 77.8 | 35.7 | 34.5 | N/A | 6.738 | 36 | 6473f9996d758fde48a181f37cc5de575aff1606 | Neko-Institute-of-Science/pygmalion-7b | ||
49.7 | 51 | 77.8 | 35.7 | 34.3 | N/A | other | 7 | 1 | afb4604a06c8541960fb51240259777764c4ce7e | TheBloke/Planner-7B-fp16 | |
49.7 | 46.2 | 71.9 | 40.7 | 39.9 | finetuned | other | 13 | 16 | f17a52b8067d551a814069d2c710e1f5c487a3ce | shibing624/chinese-llama-plus-13b-hf | |
49.7 | 51 | 77.8 | 35.7 | 34.3 | pretrained | other | 6.738 | 0 | f356572651e58fb337d610470d4b36976e7fb802 | huggingface/llama-7b | |
49.6 | 42.1 | 68 | 49.3 | 38.9 | N/A | 6 | 1 | 41bd1937dbc51f9e589d310bddab5b4c1409e783 | togethercomputer/GPT-JT-6B-v0 | ||
49.5 | 47.9 | 73.1 | 35.4 | 41.5 | N/A | other | 7 | 307 | 14c23f9fa775ab5ce49010418f00df06d92b0b13 | ehartford/WizardLM-7B-Uncensored | |
49.4 | 49.2 | 70.5 | 38.4 | 39.7 | N/A | other | 7 | 43 | 0deb5a13732f1e3e3240ea83f403c57283fe2dc8 | shibing624/chinese-alpaca-plus-7b-hf | |
48.9 | 50.3 | 77.9 | 32.3 | 35.1 | N/A | cc-by-sa-3.0 | 7 | 420 | 925e0d80e50e77aaddaf9c3ced41ca4ea23a1025 | mosaicml/mpt-7b-instruct | |
48.7 | 50.2 | 77.8 | 31.9 | 35 | N/A | cc-by-sa-3.0 | 7 | 420 | 1fc4634127ec64a45716003578b9cfae23265849 | mosaicml/mpt-7b-instruct | |
48.6 | 47.4 | 61.2 | 47.6 | 38 | N/A | cc-by-nc-4.0 | 30 | 30 | 80bd55898b06c7c363c467dec877b8b32702a2c4 | facebook/galactica-30b | |
48.2 | 49.6 | 58.2 | 43.8 | 41.2 | N/A | cc-by-nc-4.0 | 30 | 54 | a1f0c4bedd65b485a0d4d3a3bd60d7a4599f1eaf | GeorgiaTechResearchInstitute/galpaca-30b | |
48.2 | 38.8 | 59 | 46.7 | 48.1 | N/A | 6 | 1209 | 162b620e3078b03eefff94eb5f762d4093425fb5 | THUDM/chatglm2-6b | ||
48.2 | 43.7 | 72.2 | 41.3 | 35.5 | N/A | apache-2.0 | 7 | 57 | e5961def23172a2384543940e773ab676033c963 | openlm-research/open_llama_7b_v2 | |
48.1 | 40.9 | 67.1 | 47.2 | 37.1 | N/A | apache-2.0 | 6 | 294 | f34aa35f906895602c1f86f5685e598afdea8051 | togethercomputer/GPT-JT-6B-v1 | |
48 | 47.1 | 73.6 | 25.5 | 46 | N/A | other | 7 | 17 | d102fe3b68f1a5a50d547e4fd1c8b33b783c993b | TheBloke/koala-7B-HF | |
47.4 | 40.7 | 69 | 43.6 | 36.2 | N/A | 7 | 698 | 8baef65be8363f3b5670adfe9a0b9c0311962d90 | baichuan-inc/Baichuan-7B | ||
47 | 47.9 | 78.1 | 27.8 | 34.3 | pretrained | apache-2.0 | 7 | 687 | 378337427557d1df3e742264a2901a49f25d4eb1 | tiiuae/falcon-7b | |
46.9 | 44.1 | 72 | 37.6 | 34 | N/A | apache-2.0 | 7 | 98 | 95667a602ff2646bf67fe3a57c4eb9a1edec87fe | togethercomputer/RedPajama-INCITE-7B-Instruct | |
46.9 | 44.1 | 72 | 37.6 | 34 | N/A | apache-2.0 | 7 | 98 | 95667a602ff2646bf67fe3a57c4eb9a1edec87fe | togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1 | |
46.7 | 48 | 72.8 | 26 | 39.9 | N/A | apache-2.0 | 20 | 4 | 1a5b8d25587eab67d837621a6c9423e7ef6df289 | h2oai/h2ogpt-gm-oasst1-en-1024-20b | |
46.6 | 48.8 | 74.4 | 26.2 | 36.9 | N/A | 20 | 0 | 20b347273d90da7c2c9eb4c32d4173dba862a0d2 | dvruette/gpt-neox-20b-full-precision | ||
46.3 | 48.5 | 74.6 | 26.4 | 35.6 | finetuned | 20 | 0 | 4aec11ef19103796fb21387ce925b63c9d61dae1 | dvruette/oasst-gpt-neox-20b-1000-steps | ||
46.2 | 46.3 | 76.2 | 27 | 35.4 | pretrained | other | 66 | 170 | 7259969061237fe940036d22bea0fd349e4485e9 | facebook/opt-66b | |
46.2 | 46.8 | 71.9 | 32.4 | 33.9 | pretrained | bsd-3-clause | 16 | 15 | b65951b0cf7c5639f73caea801a892788608ed69 | Salesforce/codegen-16B-nl | |
46.2 | 39.8 | 70.3 | 35.2 | 39.5 | N/A | cc-by-sa-3.0 | 7 | 11 | b8fbe09571a71603ab517fe897a1281005060b62 | Vmware/open-llama-7b-v2-open-instruct | |
46.2 | 45.6 | 74.1 | 28.8 | 36.1 | N/A | apache-2.0 | 7 | 649 | a5e85ae1941e31bb705adbcafce9b0dfd6f3a48b | mosaicml/mpt-7b-storywriter | |
46 | 45.6 | 74 | 29.9 | 34.5 | N/A | apache-2.0 | 20 | 690 | d386708e84d862a65f7d2b4989f64750cb657227 | togethercomputer/GPT-NeoXT-Chat-Base-20B | |
45.8 | 46.9 | 72.8 | 26.2 | 37.5 | finetuned | apache-2.0 | 20 | 36 | 3bdf6f870ca14bcc5587b666fbe57488f7854d30 | h2oai/h2ogpt-oasst1-512-20b | |
45.7 | 45.7 | 74.2 | 26.7 | 36.1 | N/A | apache-2.0 | 7 | 649 | a5e85ae1941e31bb705adbcafce9b0dfd6f3a48b | mosaicml/mpt-7b-storywriter | |
45.4 | 46.4 | 70 | 26.2 | 39.2 | N/A | apache-2.0 | 12 | 279 | 293df535fe7711a5726987fc2f17dfc87de452a1 | OpenAssistant/oasst-sft-1-pythia-12b | |
45.3 | 43.9 | 65.2 | 30 | 42 | N/A | cc-by-nc-sa-4.0 | 7 | 16 | 6ed0dca683685cb5b9e7df599f87d311f00ba6db | psmathur/orca_mini_7b | |
45.2 | 47.4 | 72.6 | 26.4 | 34.4 | N/A | apache-2.0 | 20 | 7 | b3a6bf4250a037c09e451344e2a4e987011b79de | h2oai/h2ogpt-gm-oasst1-multilang-1024-20b | |
45.1 | 46.7 | 67.7 | 28.5 | 37.6 | N/A | cc | 7 | 4 | 75741b55ad462330e3498d1506f438f835152177 | VMware/open-llama-0.7T-7B-open-instruct-v1.1 | |
45 | 46.4 | 72.1 | 26.2 | 35.5 | N/A | 20 | 0 | f0462a8b7908f61202d86e6a9a2996d8339363b5 | dvruette/oasst-gpt-neox-20b-3000-steps | ||
44.9 | 46 | 67.8 | 30.6 | 35.2 | pretrained | apache-2.0 | 7 | 91 | 6fb184ff23774c25bf84b3628e49c8b78372c7be | openlm-research/open_llama_7b | |
44.8 | 45.3 | 67.7 | 27.8 | 38.2 | N/A | 12 | 0 | c21fbece4253841f2d6e15f04f60fe1ba6f990dd | dvruette/oasst-pythia-12b-pretrained-sft | ||
44.8 | 43.2 | 74.1 | 26.6 | 35.2 | N/A | other | 30 | 130 | ceea0a90ac0f6fae7c2c34bcb40477438c152546 | facebook/opt-30b | |
44.7 | 45.7 | 68.6 | 26.8 | 37.8 | finetuned | apache-2.0 | 12 | 318 | 626b8c140cfdedb119dfb78c626cd772283dee33 | OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5 | |
44.7 | 41.2 | 66.9 | 36.5 | 34.2 | N/A | bigscience-openrail-m | 9 | 92b955a3ff74aa577fa0d8517dfc314847ef60af | digitous/GPT-R | ||
44.6 | 46.2 | 71.6 | 27.7 | 33 | pretrained | apache-2.0 | 7 | 76 | 78f7e482443971f4873ba3239f0ac810a367833b | togethercomputer/RedPajama-INCITE-Base-7B-v0.1 | |
44.6 | 46.2 | 71.6 | 27.7 | 33 | N/A | apache-2.0 | 7 | 76 | 78f7e482443971f4873ba3239f0ac810a367833b | togethercomputer/RedPajama-INCITE-7B-Base | |
44.4 | 44 | 70.3 | 26.6 | 36.5 | N/A | apache-2.0 | 12 | 21 | 275c9b71bfab4e271d1ed85515c61e317b6ef65e | OpenAssistant/pythia-12b-sft-v8-7k-steps | |
44.3 | 45.5 | 72.8 | 26.8 | 32.1 | N/A | apache-2.0 | 20 | 63 | 1a80940a290452af71caf17a8e520955eb338e0f | KoboldAI/GPT-NeoX-20B-Erebus | |
44.2 | 37.9 | 63.3 | 32.8 | 42.6 | finetuned | creativeml-openrail-m | 1 | 79d5a4d53953ca1c26bc2155f168b7e2108f377f | AlekseyKorshuk/chatml-pyg-v1 | ||
44.2 | 43.1 | 69.7 | 25.9 | 38 | finetuned | apache-2.0 | 12 | 5 | e547fffafb382fd39ef5de35ba3b5afc1b43e74d | h2oai/h2ogpt-gm-oasst1-en-1024-12b | |
44.1 | 42.3 | 70.1 | 27.4 | 36.8 | N/A | apache-2.0 | 12 | 0 | 142e306db8e279a07c557ea5a919ab7e7a4af17c | OpenAssistant/pythia-12b-sft-v8-2.5k-steps | |
44 | 42.1 | 70.8 | 26.9 | 36.1 | N/A | apache-2.0 | 7 | 83 | 47b94a739e2f3164b438501c8684acc5d5acc146 | togethercomputer/RedPajama-INCITE-Chat-7B-v0.1 | |
44 | 42.1 | 70.8 | 26.9 | 36.1 | N/A | apache-2.0 | 7 | 83 | 47b94a739e2f3164b438501c8684acc5d5acc146 | togethercomputer/RedPajama-INCITE-7B-Chat | |
43.9 | 41.6 | 69 | 30.7 | 34.5 | N/A | creativeml-openrail-m | 2 | 4c4a5caf5d9049a47f5565b72e5a53dede08ac8b | digitous/Javelin-R | ||
43.9 | 45.7 | 73.5 | 25 | 31.6 | pretrained | apache-2.0 | 20.739 | 409 | 9369f145ca7b66ef62760f9351af951b2d53b77f | EleutherAI/gpt-neox-20b | |
43.9 | 42.7 | 70.5 | 26.2 | 36.1 | finetuned | creativeml-openrail-m | 4 | bee7068ab002784420a1a30170db3906185359f2 | digitous/Javelin-GPTJ | ||
43.8 | 45 | 72.7 | 26 | 31.6 | N/A | apache-2.0 | 20 | 9 | dd98d514b5aff4e820922c88a73d6d5bf17f332e | KoboldAI/GPT-NeoX-20B-Skein | |
43.7 | 42.3 | 70.2 | 26 | 36.4 | N/A | apache-2.0 | 12 | 25 | c6bb0fe363e0105839d34ca757793b61c9606f95 | h2oai/h2ogpt-oasst1-512-12b | |
43.7 | 41.7 | 68 | 30.8 | 34.4 | finetuned | creativeml-openrail-m | 5 | b881231ab6ea85da2a9a139f282df85d1d18b002 | digitous/Javalion-R | ||
43.7 | 42.6 | 68.7 | 24.9 | 38.7 | N/A | 6 | 10 | acfe27303f74129930fef5e6fadbc5f58c6b8590 | KoboldAI/GPT-J-6B-Skein | ||
43.7 | 42.4 | 72.5 | 25.9 | 33.8 | finetuned | mit | 12 | 1827 | 19308160448536e378e3db21a73a751579ee7fdd | databricks/dolly-v2-12b | |
43.6 | 41.4 | 67.7 | 28.5 | 36.9 | N/A | apache-2.0 | 6 | 3 | 6413b1d9e8b58df9d3aac91a862e8d505d8c6716 | TehVenom/Dolly_Shygmalion-6b-Dev_V8P2 | |
43.6 | 44.5 | 69.6 | 25.2 | 34.9 | finetuned | mit | 7 | 119 | d632f0c8b75b1ae5b26b250d25bfba4e99cb7c6f | databricks/dolly-v2-7b | |
43.5 | 42.5 | 69.2 | 25.4 | 36.9 | N/A | 0 | 4fbfe9eae03a1d6ecf60fda8cf39c4123f0438bd | digitous/Adventien-GPTJ | |||
43.4 | 41.5 | 68.8 | 26.6 | 36.8 | N/A | apache-2.0 | 12 | 6 | 37ca702e957a4b740689d67c58c284224e2fbae2 | OpenAssistant/pythia-12b-pre-v8-12.5k-steps | |
43.4 | 40.4 | 67.4 | 31.2 | 34.5 | N/A | creativeml-openrail-m | 1 | f6963f77098d8421ff4a1cf4d36f1e94c6c8f44b | digitous/Janin-R | ||
43.3 | 43.8 | 69.2 | 25.4 | 34.7 | N/A | creativeml-openrail-m | 0 | 4dff006b2ea7e8d9b067dfe8af8ca1a16bc44dce | digitous/Skegma-GPTJ | ||
43.3 | 41.3 | 62.4 | 27.6 | 42 | N/A | apache-2.0 | 7 | 3 | 29604e6e19822531b0d49d3f19abef603a97d0ec | h2oai/h2ogpt-gm-oasst1-en-1024-open-llama-7b-preview-400bt | |
43.2 | 42.3 | 68.9 | 26 | 35.6 | finetuned | 13 | 8 | 7faeb395c26189eeab9bf3a98994696687ad31a3 | Pirr/pythia-13b-deduped-green_devil | ||
43.2 | 41.9 | 68.7 | 26.8 | 35.4 | N/A | creativeml-openrail-m | 1 | 3ce176bc0f91cae416c78e99f964f54b12472de0 | digitous/Javalion-GPTJ | ||
43.2 | 40.7 | 67 | 29.3 | 35.6 | N/A | 6 | 3 | fa3d503bca50c947e7a5bbde4bdd82f699f65c02 | TehVenom/PPO_Shygmalion-V8p4_Dev-6b | ||
43.1 | 41.6 | 64.2 | 26.3 | 40.4 | N/A | 9 | 0 | 0e201b6f344ac6382dda40d389e1c9144a87d027 | dvruette/oasst-pythia-6.9b-4000-steps | ||
43.1 | 41.6 | 61.5 | 26.8 | 42.4 | N/A | cc-by-nc-sa-4.0 | 3 | 98 | fd2754e80ce80757a3a68a840d7d287dd7def676 | psmathur/orca_mini_3b | |
43 | 38.9 | 64.8 | 28 | 40.4 | finetuned | creativeml-openrail-m | 6 | 589 | 30e2405100eac6bd53f75964cc7345eeafd19f7d | PygmalionAI/pygmalion-6b | |
43 | 40.2 | 66.4 | 30.4 | 34.8 | N/A | bigscience-openrail-m | 6 | 1 | 930dc82245c607ce43558a0e6c0225e77b341ea6 | TehVenom/GPT-J-Pyg_PPO-6B-Dev-V8p4 | |
43 | 41.9 | 68.5 | 27.6 | 33.9 | N/A | 6 | 14 | 108fabf8a916900525492c294c50998d7c09f10b | TehVenom/Dolly_Shygmalion-6b | ||
43 | 41.3 | 66 | 26.8 | 37.9 | N/A | 6 | 1 | 83d8c754aac12f838d7c847d4352a09396c383d0 | Corianas/gpt-j-6B-Dolly | ||
43 | 40.9 | 67.3 | 27.4 | 36.2 | N/A | creativeml-openrail-m | 0 | a6773861798f2abea3849514aa6f60961518af9c | digitous/Janin-GPTJ | ||
43 | 40.4 | 67.1 | 29.3 | 35.3 | N/A | apache-2.0 | 6 | 8 | f30709dba36c665869f9ac8cd0cef5a8a2e7c8df | TehVenom/PPO_Pygway-V8p4_Dev-6b | |
42.9 | 41.4 | 67.6 | 26.6 | 36 | pretrained | apache-2.0 | 6 | 1214 | 47e169305d2e8376be1d31e765533382721b2cc1 | EleutherAI/gpt-j-6b | |
42.9 | 39.8 | 67.1 | 27.7 | 36.9 | N/A | mit | 6 | 13 | afa5a11b24cb23eee708e17c83b920a788e9e07b | KoboldAI/GPT-J-6B-Shinen | |
42.8 | 40.9 | 67.1 | 27.4 | 35.7 | N/A | mit | 6 | 8 | 036bb03496d648ddc8cf932ad91df8ef1287116c | KoboldAI/GPT-J-6B-Janeway | |
42.8 | 41.6 | 66.2 | 27.3 | 36.1 | N/A | cc | 2.91 | 2 | c588a5924749b86a6cb36a687dafa544c189bb6f | Fredithefish/RedPajama-INCITE-Chat-3B-Instruction-Tuning-with-GPT-4 | |
42.8 | 42.3 | 68.6 | 25.9 | 34.5 | N/A | bsd-3-clause | 6 | 4 | dff91c0aea702edbea3528344d01d8b9aaee6e39 | Salesforce/codegen-6B-nl | |
42.8 | 41.9 | 68.3 | 27.3 | 33.9 | N/A | 9 | 2667b0e0b705ed23f81f3e2b69673d722e8f4964 | TehVenom/ChanMalion | |||
42.8 | 42.8 | 67.6 | 26.2 | 34.4 | N/A | apache-2.0 | 3 | 104 | f0e0995eba801096ed04cb87931d96a8316871af | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | |
42.8 | 42.8 | 68.4 | 27.1 | 33 | N/A | 6 | 1 | f239eb8d24fe26db3b0a9a69115dc305fc9351af | TehVenom/Dolly_Malion-6b | ||
42.7 | 40 | 68.7 | 27.4 | 34.6 | N/A | apache-2.0 | 7 | 60 | 97aa918c383820e1a69f042801091d7deb996c20 | togethercomputer/Pythia-Chat-Base-7B | |
42.6 | 39.4 | 71.3 | 26.9 | 32.7 | N/A | 7 | 2 | d62d83b8eb7a6ba012a762752a5b5679add3b40c | KoboldAI/fairseq-dense-6.7B | ||
42.6 | 41.8 | 67.8 | 28.4 | 32.5 | N/A | apache-2.0 | 6 | 18 | b31d25819e00d5031ccdb22a9584f0850dcfe39c | KoboldAI/PPO_Pygway-6b-Mix | |
42.6 | 40 | 70.1 | 25.3 | 34.9 | N/A | other | 13 | 115 | 8a949353677d2b971910a6c4afcc70e95d838c2a | KoboldAI/OPT-13B-Erebus | |
42.6 | 41.4 | 70.3 | 25.6 |