finetuning-llms
Here are 185 public repositories matching this topic...
Mastering Applied AI, One Concept at a Time
-
Updated
May 23, 2025 - Jupyter Notebook
End to End Generative AI Industry Projects on LLM Models with Deployment_Awesome LLM Projects
-
Updated
Jan 24, 2025
Auto Data is a library designed for quick and effortless creation of datasets tailored for fine-tuning Large Language Models (LLMs).
-
Updated
Oct 31, 2024 - Python
🚀 Easy, open-source LLM finetuning with one-line commands, seamless cloud integration, and popular optimization frameworks. ✨
-
Updated
Aug 14, 2024 - Python
On Memorization of Large Language Models in Logical Reasoning
-
Updated
Mar 29, 2025 - Python
Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"
-
Updated
Mar 23, 2025 - Jupyter Notebook
MediNotes: SOAP Note Generation through Ambient Listening, Large Language Model Fine-Tuning, and RAG
-
Updated
Dec 18, 2024 - Python
A Gradio web UI for Large Language Models. Supports LoRA/QLoRA finetuning,RAG(Retrieval-augmented generation) and Chat
-
Updated
Nov 26, 2023 - Python
Fine-tune Mistral 7B to generate fashion style suggestions
-
Updated
Jan 2, 2024 - Python
the small distributed language model toolkit; fine-tune state-of-the-art LLMs anywhere, rapidly
-
Updated
Oct 19, 2024 - Python
[NeurIPS 2023] Make Your Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning
-
Updated
Jun 2, 2023 - Python
Medical Language Model fine-tuned using pretraining, instruction tuning, and Direct Preference Optimization (DPO). Progresses from general medical knowledge to specific instruction following, with experiments in preference alignment for improved medical text generation and understanding.
-
Updated
Oct 4, 2024 - Jupyter Notebook
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
-
Updated
Mar 7, 2024 - Python
SEIKO is a novel reinforcement learning method to efficiently fine-tune diffusion models in an online setting. Our methods outperform all baselines (PPO, classifier-based guidance, direct reward backpropagation) for fine-tuning Stable Diffusion.
-
Updated
Jul 18, 2024 - Python
A open-source framework designed to adapt pre-trained Language Models (LLMs), such as Llama, Mistral, and Mixtral, to a wide array of domains and languages.
-
Updated
May 27, 2024 - Python
[EMNLP 2024] Quantize LLM to extremely low-bit, and finetune the quantized LLMs
-
Updated
Jul 18, 2024 - Python
Fine-Tuning LLMs (Gemma, LLaMA, Mistral, etc.) A practical guide to fine-tuning various large language models using popular frameworks. Includes examples, scripts, and tips for efficient training on custom datasets.
-
Updated
Jun 2, 2025 - Jupyter Notebook
A curated list of Parameter Efficient Fine-tuning papers with a TL;DR
-
Updated
Sep 19, 2024
IndexTTS Fine-tuning notebooks
-
Updated
Jun 17, 2025 - Jupyter Notebook
Improve this page
Add a description, image, and links to the finetuning-llms topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the finetuning-llms topic, visit your repo's landing page and select "manage topics."