dpo
Here are 41 public repositories matching this topic...
tensorflow를 사용하여 텍스트 전처리부터, Topic Models, BERT, GPT, LLM과 같은 최신 모델의 다운스트림 태스크들을 정리한 Deep Learning NLP 저장소입니다.
-
Updated
Feb 22, 2024 - Jupyter Notebook
Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first approach
-
Updated
Jan 15, 2024 - Python
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
-
Updated
Apr 25, 2024 - Python
CodeUltraFeedback for aligning large language models to coding preferences
-
Updated
Mar 17, 2024 - Python
Various training, inference and validation code and results related to Open LLM's that were pretrained (full or partially) on the Dutch language.
-
Updated
Apr 9, 2024 - Jupyter Notebook
A Laravel package to simplify using DPO Payment API in your application. https://dpogroup.com
-
Updated
Sep 8, 2023 - PHP
Data and models for the paper "Configurable Safety Tuning of Language Models with Synthetic Preference Data"
-
Updated
Apr 23, 2024 - Python
A open-source framework designed to adapt pre-trained Language Models (LLMs), such as Llama, Mistral, and Mixtral, to a wide array of domains and languages.
-
Updated
Apr 25, 2024 - Python
This is the DPO Pay plugin for WooCommerce.
-
Updated
Jan 4, 2024 - PHP
EasyRLHF aims to provide an easy and minimal interface to train aligned language models, using off-the-shelf solutions and datasets
-
Updated
Dec 12, 2023 - Python
Examples for using the SiLLM framework for training and running Large Language Models (LLMs) on Apple Silicon
-
Updated
Apr 23, 2024 - Python
Improve this page
Add a description, image, and links to the dpo topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the dpo topic, visit your repo's landing page and select "manage topics."