Skip to content

AI-App/OpenDevin.OpenDevin

Β 
Β 

Repository files navigation

Contributors Forks Stargazers Issues MIT License
Join our Slack community Join our Discord community
Logo

OpenDevin: Code Less, Make More

πŸ—‚οΈ Table of Contents
  1. 🎯 Mission
  2. πŸ€” What is Devin?
  3. 🐚 Why OpenDevin?
  4. 🚧 Project Status
  5. πŸš€ Get Started
  6. ⭐️ Research Strategy
  7. 🀝 How to Contribute
  8. πŸ€– Join Our Community
  9. πŸ› οΈ Built With
  10. πŸ“œ License

🎯 Mission

OpenDevin.webm

Welcome to OpenDevin, an open-source project aiming to replicate Devin, an autonomous AI software engineer who is capable of executing complex engineering tasks and collaborating actively with users on software development projects. This project aspires to replicate, enhance, and innovate upon Devin through the power of the open-source community.

↑ Back to Top ↑

πŸ€” What is Devin?

Devin represents a cutting-edge autonomous agent designed to navigate the complexities of software engineering. It leverages a combination of tools such as a shell, code editor, and web browser, showcasing the untapped potential of LLMs in software development. Our goal is to explore and expand upon Devin's capabilities, identifying both its strengths and areas for improvement, to guide the progress of open code models.

↑ Back to Top ↑

🐚 Why OpenDevin?

The OpenDevin project is born out of a desire to replicate, enhance, and innovate beyond the original Devin model. By engaging the open-source community, we aim to tackle the challenges faced by Code LLMs in practical scenarios, producing works that significantly contribute to the community and pave the way for future advancements.

↑ Back to Top ↑

🚧 Project Status

OpenDevin is currently a work in progress, but you can already run the alpha version to see the end-to-end system in action. The project team is actively working on the following key milestones:

  • UI: Developing a user-friendly interface, including a chat interface, a shell demonstrating commands, and a web browser.
  • Architecture: Building a stable agent framework with a robust backend that can read, write, and run simple commands.
  • Agent Capabilities: Enhancing the agent's abilities to generate bash scripts, run tests, and perform other software engineering tasks.
  • Evaluation: Establishing a minimal evaluation pipeline that is consistent with Devin's evaluation criteria.

After completing the MVP, the team will focus on research in various areas, including foundation models, specialist capabilities, evaluation, and agent studies.

↑ Back to Top ↑

⚠️ Caveats and Warnings

  • OpenDevin is still an alpha project. It is changing very quickly and is unstable. We are working on getting a stable release out in the coming weeks.
  • OpenDevin will issue many prompts to the LLM you configure. Most of these LLMs cost money--be sure to set spending limits and monitor usage.
  • OpenDevin runs bash commands within a Docker sandbox, so it should not affect your machine. But your workspace directory will be attached to that sandbox, and files in the directory may be modified or deleted.
  • Our default Agent is currently the MonologueAgent, which has limited capabilities, but is fairly stable. We're working on other Agent implementations, including SWE Agent. You can read about our current set of agents here.

πŸš€ Get Started

The easiest way to run OpenDevin is inside a Docker container. You can run:

# Your OpenAI API key, or any other LLM API key
export LLM_API_KEY="sk-..."

# The directory you want OpenDevin to modify. MUST be an absolute path!
export WORKSPACE_DIR=$(pwd)/workspace

docker run \
    -e LLM_API_KEY \
    -e WORKSPACE_MOUNT_PATH=$WORKSPACE_DIR \
    -v $WORKSPACE_DIR:/opt/workspace_base \
    -v /var/run/docker.sock:/var/run/docker.sock \
    -p 3000:3000 \
    ghcr.io/opendevin/opendevin:main

Replace $(pwd)/workspace with the path to the code you want OpenDevin to work with.

You can find opendevin running at http://localhost:3000.

See Development.md for instructions on running OpenDevin without Docker.

πŸ€– LLM Backends

OpenDevin can work with any LLM backend. For a full list of the LM providers and models available, please consult the litellm documentation.

The LLM_MODEL environment variable controls which model is used in programmatic interactions, but choosing a model in the OpenDevin UI will override this setting.

The following environment variables might be necessary for some LLMs:

  • LLM_API_KEY
  • LLM_BASE_URL
  • LLM_EMBEDDING_MODEL
  • LLM_DEPLOYMENT_NAME
  • LLM_API_VERSION

Note on Alternative Models: Some alternative models may prove more challenging to tame than others. Fear not, brave adventurer! We shall soon unveil LLM-specific documentation to guide you on your quest. And if you've already mastered the art of wielding a model other than OpenAI's GPT, we encourage you to share your setup instructions with us.

There is also documentation for running with local models using ollama.

⭐️ Research Strategy

Achieving full replication of production-grade applications with LLMs is a complex endeavor. Our strategy involves:

  1. Core Technical Research: Focusing on foundational research to understand and improve the technical aspects of code generation and handling.
  2. Specialist Abilities: Enhancing the effectiveness of core components through data curation, training methods, and more.
  3. Task Planning: Developing capabilities for bug detection, codebase management, and optimization.
  4. Evaluation: Establishing comprehensive evaluation metrics to better understand and improve our models.

↑ Back to Top ↑

🀝 How to Contribute

OpenDevin is a community-driven project, and we welcome contributions from everyone. Whether you're a developer, a researcher, or simply enthusiastic about advancing the field of software engineering with AI, there are many ways to get involved:

  • Code Contributions: Help us develop the core functionalities, frontend interface, or sandboxing solutions.
  • Research and Evaluation: Contribute to our understanding of LLMs in software engineering, participate in evaluating the models, or suggest improvements.
  • Feedback and Testing: Use the OpenDevin toolset, report bugs, suggest features, or provide feedback on usability.

For details, please check this document.

↑ Back to Top ↑

πŸ€– Join Our Community

Now we have both Slack workspace for the collaboration on building OpenDevin and Discord server for discussion about anything related, e.g., this project, LLM, agent, etc.

If you would love to contribute, feel free to join our community (note that now there is no need to fill in the form). Let's simplify software engineering together!

🐚 Code less, make more with OpenDevin.

Star History Chart

πŸ› οΈ Built With

OpenDevin is built using a combination of powerful frameworks and libraries, providing a robust foundation for its development. Here are the key technologies used in the project:

FastAPI uvicorn LiteLLM Docker Ruff MyPy LlamaIndex React

Please note that the selection of these technologies is in progress, and additional technologies may be added or existing ones may be removed as the project evolves. We strive to adopt the most suitable and efficient tools to enhance the capabilities of OpenDevin.

↑ Back to Top ↑

πŸ“œ License

Distributed under the MIT License. See LICENSE for more information.

↑ Back to Top ↑

About

🐚 OpenDevin: Code Less, Make More

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 62.1%
  • TypeScript 30.0%
  • Makefile 2.7%
  • Shell 2.4%
  • JavaScript 0.9%
  • Dockerfile 0.9%
  • Other 1.0%