Skip to content

Darren-greenhand/Little_Camel

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Start from Alpaca

Novices also want to join the bandwagon of building LLMs !

Start from Organizing and assembling the Existing repository, and then try to add something new !!

Source

CheckPoints:

Building the model:

Stanford_alpaca

【The 1️⃣first popular player using the LLaMA】This is the repo for the Stanford Alpaca project, which aims to build and share an instruction-following LLaMA model. The repo contains:

Alpaca-lora

This repository contains code for reproducing the Stanford Alpaca results using low-rank adaptation (LoRA). We provide an Instruct model of similar quality to text-davinci-003 that can run on a Raspberry Pi (for research), and the code is easily extended to the 13b, 30b, and 65b models.

In addition to the training code, which runs within hours on a single RTX 4090, we publish a script for downloading and inference on the foundation model and LoRA, as well as the resulting LoRA weights themselves. To fine-tune cheaply and efficiently, we use Hugging Face's PEFT as well as Tim Dettmers' bitsandbytes.

Without hyperparameter tuning, the LoRA model produces outputs comparable to the Stanford Alpaca model. (Please see the outputs included below.) Further tuning might be able to achieve better performance

BELLE: Be Everyone's Large Language model Engine

项目包含以下内容:

项目包含以下内容:

Chinese-LLaMA-Alpaca

本项目开源了中文LLaMA模型和经过指令精调的Alpaca大模型。这些模型在原版LLaMA的基础上扩充了中文词表并使用了中文数据进行二次预训练,进一步提升了中文基础语义理解能力。同时,在中文LLaMA的基础上,本项目使用了中文指令数据进行指令精调,显著提升了模型对指令的理解和执行能力。

主要内容:

🚀 开源了经过中文文本数据预训练的中文LLaMA大模型

🚀 开源了进一步经过指令精调的中文Alpaca大模型

🚀 快速地使用笔记本电脑(个人PC)本地部署和体验量化版大模型

llama.cpp

Inference of LLaMA model in pure C/C++ 😩

The main goal is to run the model using 4-bit quantization on a MacBook(also supports Windows)

  • Plain C/C++ implementation without dependencies
  • Apple silicon first-class citizen - optimized via ARM NEON and Accelerate framework
  • AVX2 support for x86 architectures
  • Mixed F16 / F32 precision
  • 4-bit quantization support
  • Runs on the CPU

Data && Deployment

AlpacaDataCleaned

This repository hosts a cleaned and curated version of a dataset used to train the Alpaca LLM (Large Language Model). The original dataset had several issues that are addressed in this cleaned version.

Alpaca-LoRA-Serve

Alpaca-LoRA as a Chatbot Service

img

Alpaca.cpp

Run a fast ChatGPT-like model locally on your device.

Based on the Alpaca-lora and the horiible llama.cpp but add a chat interface.

免责声明

本项目相关资源仅供学术研究之用,严禁用于商业用途。使用涉及第三方代码的部分时,请严格遵循相应的开源协议。模型生成的内容受模型计算、随机性和量化精度损失等因素影响,本项目无法对其准确性作出保证。对于模型输出的任何内容,本项目不承担任何法律责任,亦不对因使用相关资源和输出结果而可能产生的任何损失承担责任。

本项目由个人及协作者业余时间发起并维护,因此无法保证能及时回复解决相应问题。

About

Well, it develops too fast and I can't follow it

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages