Skip to content

KennethanCeyer/awesome-llmops

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

56 Commits
 
 
 
 
 
 

Repository files navigation

Awesome LLMOps

Awesome

Awesome LLMOps - Awesome list of LLMOps

Table of Contents

What is LLMOps?

LLMOps is a part of MLOps practices, specialized form of MLOps that focuses on managing the entire lifecycle of large language models(LLM).

Starting in 2021, as LLMs evolved rapidly and the technology matured, we began to focus on practices for managing LLMs efficiently, and LLMOps, which are adaptations of traditional MLOps practices to LLMs, began to be talked about.

LLMOps vs MLOps

LLMOps MLOps
Definition Tools and infrastructure specifically for the development and deployment of large language models Tools and infrastructure for general machine learning workflows
Focus Unique requirements and challenges of large language models General machine learning workflows
Key technologies Language model, Transformers library, human-in-the-loop annotation platforms Kubeflow, MLflow, TensorFlow Extended
Key skills NLP expertise, knowledge of large language models, data management for text data Data engineering, DevOps, Software engineering, Machine learning expertise
Key challenges Managing and labeling large amounts of text data, fine-tuning foundation models for specific tasks, ensuring fairness and ethics in language models Managing complex data pipelines, ensuring model interpretability and explainability, addressing model bias and fairness
Industry adoption Emerging, with a growing number of startups and companies focusing on LLMOps Established, with a large ecosystem of tools and frameworks available
Future outlook LLMOps is expected to become an increasingly important area of study as large language models become more prevalent and powerful MLOps will continue to be a critical component of the machine learning industry, with a focus on improving efficiency, scalability, and model reliability

⬆️ Go to top

Prompt Engineering

⬆️ Go to top

Models

Name Parameter size Announcement date
BERT-Large (336M) 336 million 2018
T5 (11B) 11 billion 2020
Gopher (280B) 280 billion 2021
GPT-J (6B) 6 billion 2021
LaMDA (137B) 137 billion 2021
Megatron-Turing NLG (530B) 530 billion 2021
T0 (11B) 11 billion 2021
Macaw (11B) 11 billion 2021
GLaM (1.2T) 1.2 trillion 2021
T5 FLAN (540B) 540 billion 2022
OPT-175B (175B) 175 billion 2022
ChatGPT (175B) 175 billion 2022
GPT 3.5 (175B) 175 billion 2022
AlexaTM (20B) 20 billion 2022
Bloom (176B) 176 billion 2022
Bard Not yet announced 2023
GPT 4 Not yet announced 2023
AlphaCode (41.4B) 41.4 billion 2022
Chinchilla (70B) 70 billion 2022
Sparrow (70B) 70 billion 2022
PaLM (540B) 540 billion 2022
NLLB (54.5B) 54.5 billion 2022
Alexa TM (20B) 20 billion 2022
Galactica (120B) 120 billion 2022
UL2 (20B) 20 billion 2022
Jurassic-1 (178B) 178 billion 2022
LLaMA (65B) 65 billion 2023
Stanford Alpaca (7B) 7 billion 2023
GPT-NeoX 2.0 (20B) 20 billion 2023
BloombergGPT 50 billion 2023
Dolly 6 billion 2023
Jurassic-2 Not yet announced 2023
OpenAssistant LLaMa 30 billion 2023
Koala 13 billion 2023
Vicuna 13 billion 2023
PaLM2 Not yet announced, Smaller than PaLM1 2023
LIMA 65 billion 2023
MPT 7 billion 2023
Falcon 40 billion 2023
Llama 2 70 billion 2023
Google Gemini Not yet announced 2023
Microsoft Phi-2 2.7 billion 2023
Grok-0 33 billion 2023
Grok-1 314 billion 2023
Solar 10.7 billion 2024
Gemma 7 billion 2024
Grok-1.5 Not yet announced 2024
DBRX 132 billion 2024
Claude 3 Not yet announced 2024
Gemma 1.1 7 billion 2024
Llama 3 70 billion 2024

⬆️ Go to top

Optimization

⬆️ Go to top

Tools (GitHub)

  • Stanford Alpaca - Repo stars of tatsu-lab/stanford_alpaca - A repository of Stanford Alpaca project, a model fine-tuned from the LLaMA 7B model on 52K instruction-following demonstrations.
  • LoRA - Repo stars of microsoft/LoRA - An implementation of "LoRA: Low-Rank Adaptation of Large Language Models".
  • Dolly - Repo stars of databrickslabs/dolly - A large language model trained on the Databricks Machine Learning Platform.
  • DeepSpeed - Repo stars of microsoft/DeepSpeed - A deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
  • LMFlow - Repo stars of OptimalScale/LMFlow - An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.
  • Promptify - Repo stars of promptslab/Promptify - An utility / tookit for Prompt engineering.
  • Auto-GPT - Repo stars of Significant-Gravitas/Auto-GPT - An experimental open-source attempt to make GPT-4 fully autonomous.
  • Jarvis - Repo stars of microsoft/JARVIS - A system to connect LLMs with ML community, a composite model connector via the LLM interface.
  • dalai - Repo stars of cocktailpeanut/dalai - The cli tool to run LLaMA on the local machine.
  • haystack - Repo stars of deepset-ai/haystack -an open source NLP framework to interact with the data using Transformer models and LLMs.
  • langchain - Repo stars of hwchase17/langchain - The library which assists in the development of applications with LLM.
  • langflow - Repo stars of logspace-ai/langflow - An UI for LangChain, designed with react-flow to provide an effortless way to experiment and prototype flows.
  • deeplake - Repo stars of activeloopai/deeplake - Data Lake for Deep Learning. Build, manage, query, version, & visualize datasets.
  • alpaca-lora - Repo stars of tloen/alpaca-lora - Instruct-tune LLaMA on consumer hardware.
  • bosquet - Repo stars of BrewLLM/bosquet - LLMOps for Large Language Model based applications.
  • llama_index - Repo stars of jerryjliu/llama_index - A project that provides a central interface to connect your LLM's with external data.
  • gradio - Repo stars of gradio-app/gradio - An UI helper for the machine learning model.
  • sharegpt - Repo stars of domeccleston/sharegpt - An open-source Chrome Extension for you to share your wildest ChatGPT conversations with one click.
  • keras-nlp - Repo stars of keras-team/keras-nlp - A natural language processing library that supports users through their entire development cycle.
  • Snowkel AI - Repo stars of snorkel-team/snorkel - The data platform for foundation models.
  • promptflow - Repo stars of microsoft/promptflow - A toolkit that simplifies the development of LLM-based AI applications, from ideation to deployment.

⬆️ Go to top

Tools (Other)

  • PaLM2 API - An API service that makes PaLM2, Large Language Models (LLMs), available to Google Cloud Vertex AI.
  • Perspective API - A tool that can help mitigate toxicity and ensure healthy dialogue online.

⬆️ Go to top

RLHF

  • evals - Repo stars of openai/evals - A curated list of reinforcement learning with human feedback resources.
  • trlx - Repo stars of promptslab/Promptify - A repo for distributed training of language models with Reinforcement Learning via Human Feedback. (RLHF)
  • PaLM-rlhf-pytorch - Repo stars of lucidrains/PaLM-rlhf-pytorch - Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture.

⬆️ Go to top

Awesome

⬆️ Go to top

Contributing

We welcome contributions to the Awesome LLMOps list! If you'd like to suggest an addition or make a correction, please follow these guidelines:

  1. Fork the repository and create a new branch for your contribution.
  2. Make your changes to the README.md file.
  3. Ensure that your contribution is relevant to the topic of LLMOps.
  4. Use the following format to add your contribution:
[Name of Resource](Link to Resource) - Description of resource
  1. Add your contribution in alphabetical order within its category.
  2. Make sure that your contribution is not already listed.
  3. Provide a brief description of the resource and explain why it is relevant to LLMOps.
  4. Create a pull request with a clear title and description of your changes.

We appreciate your contributions and thank you for helping to make the Awesome LLMOps list even more awesome!

⬆️ Go to top