Skip to content

deepai-org/better-code-llm

Repository files navigation

better-code-llm

Extracting better code from large language models

Introduction

This is an experiment on techniques to extract better code from large language models. We ask the model to evaluate its own generated code.

Running:

Add your openai key to a file openai_key.txt in the root directory.

python3 main.py

Key Findings:

  • Davinci 003 seems to be even better for code than Codex?
  • The latest GPT3 is usable without ChatGPT, and better in some ways
  • GPT3 can judge the quality of its own code it wrote
  • If you ask GPT3 to explain the code it wrote, often it reveals the code isn't what you asked for.
  • Sometimes the model will generate a invalid response, but rarely!
  • We can write wacky new types of functions, eg:
    • desc_to_code(description)
    • code_to_desc(code)
    • code_quality_rating(code)
    • code_has_bugs(code)

About

Extracting better code from large language models

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages