Skip to content
/ agenta Public
forked from Agenta-AI/agenta

Easily build, version, evaluate and deploy your LLM-powered apps

License

Notifications You must be signed in to change notification settings

wdshin/agenta

 
 

Repository files navigation

37 copy

Twitter Follow Contributors Last Commit

The Open-source Developer-first LLMOps Platform

Building production-ready LLM-powered applications is currently very difficult. It involves countless iterations of prompt engineering, parameter tuning, and architectures.

Agenta provides you with the tools to quickly 🔄 iterate, 🧪 experiment, and ⚖️ evaluate your LLM apps. All without imposing any restrictions on your choice of framework, library, or model.

agenta-demo.mp4

How Agenta works:

Write your LLM-powered application as you would normally do. Feel free to use any framework, library, or model (langchain or llma_index, GPT-4 or Falcon).

Example simple application that generates baby names

    from jinja2 import Template
    import openai
    template = Template(prompt_template)
    prompt = template.render(country=country, gender=gender)

    openai.api_key = os.environ.get("OPENAI_API_KEY")  # make sure to set this manually!
    chat_completion = openai.ChatCompletion.create(
        model="gpt-3.5-turbo", messages=[{"role": "user", "content": prompt}])

    print(chat_completion.choices[0].message.content)
With two lines of code, specify the inputs and parameters for your experiment.
import agenta as ag
default_prompt = "Give me five cool names for a baby from {{country}} with this gender {{gender}}!!!!"
@ag.post
def generate(country: str, gender: str, temperature: ag.FloatParam = 0.9, prompt_template: ag.TextParam = default_prompt) -> str:
# rest of the code
Deploy your app using the Agenta CLI.
Screenshot 2023-06-19 at 15 58 34
Now your team can 🔄 iterate, 🧪 experiment, and ⚖️ evaluate different versions of your app (with your code!) in the web platform.
Screenshot 2023-06-25 at 21 08 53

Features

  • 🪄 Playground: With just a few lines of code, define the parameters and prompts you wish to experiment with. You and your team can quickly experiment and fork new versions on the web UI.

  • 📊 Version Evaluation: Define test sets, evaluate, and A/B test app versions.

  • 🚀 API Deployment Made Easy: When you are ready, deploy your LLM applications as APIs in one click.

Getting Started

Please go to docs.agenta.ai for full documentation on:

Why choose Agenta for building LLM-apps?

  • 🔨 Build quickly: You need to iterate many times on different architectures and prompts to bring apps to production. We streamline this process and allow you to do this in days instead of weeks.
  • 🏗️ Build robust apps and reduce hallucination: We provide you with the tools to systematically and easily evaluate your application to make sure you only serve robust apps to production
  • 👨‍💻 Developer-centric: We cater to complex LLM-apps and pipelines that require more than one simple prompt. We allow you to experiment and iterate on apps that have complex integration, business logic, and many prompts.
  • 🌐 Solution-Agnostic: You have the freedom to use any library and models, be it Langchain, llma_index, or a custom-written alternative.
  • 🔒 Privacy-First: We respect your privacy and do not proxy your data through third-party services. The platform and the data are hosted on your infrastructure.

Contributing

We warmly welcome contributions to Agenta. Feel free to submit issues, fork the repository, and send pull requests.

About

Easily build, version, evaluate and deploy your LLM-powered apps

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • TypeScript 53.7%
  • Python 43.7%
  • HCL 0.9%
  • Shell 0.7%
  • Dockerfile 0.7%
  • CSS 0.2%
  • JavaScript 0.1%