Skip to content

pgibler/cmdh

Repository files navigation

cmdh - Generate Linux commands using an LLM

cmdh (short for Command Helper) is a tool that invokes LLM models provided by ollama or OpenAI to convert a command request into a desired command.

Use it to look up commands and flags that that you don't know offhand or generate complex commands with chaining.

cmdh_demonstration_video.webm

Features

  • Generate Linux commands from natural language
  • Interactively run the commands using a hotkey menu system
  • Differentiates between shell command types: interactive and non-interactive
  • Supports ChatGPT, ollama, and text-generation-webui

Prerequisites

Installation

  1. Set up and configure cmdh using the following command:
git clone https://github.com/pgibler/cmdh.git && cd cmdh && ./install.sh
  1. Run it like so:
cmdh 'Output the number of lines of code committed to git last month'
  1. Interact with the result interface to run the setup commands, desired command, all of the commands, or quit.

NOTE: You will have to reload your .bashrc / .zshrc / etc. or open a new terminal to make the cmdh command available in the shell. In Debian / Ubuntu, this is done by running source ~/.bashrc.

Configuring

Before running cmdh, you will need to configure an LLM host and set configuration options required of that host.

  • Run cmdh configure to start the configuration wizard. You will be asked to select an LLM host and input settings required by that host.
  • Run cmdh configure show to display your current configuration.

OpenAI

  1. Generate an OpenAI key here.
  2. Run cmdh configure and select the OpenAI option.
  3. Select a model & input your OpenAI key.

ollama

  1. Install & run the ollama service & pull the codellama model using the following commands:
curl https://ollama.ai/install.sh | sh
ollama pull codellama
  1. Run cmdh configure, select the ollama option, and set 'codellama' as the model.

text-generation-webui

  1. Clone the repo: git clone https://github.com/oobabooga/text-generation-webui
  2. Navigate to the cloned text-generation-webui folder and start the server by running ./start_linux --api --listen.
  3. Open the web UI for text-generation-webui (http://localhost:7860), open the "Model" tab, and in the "Download model or LoRA" form, input Trelis/Llama-2-7b-chat-hf-function-calling-v2 and press "Download".
  4. Click the reload button next to the Model dropdown menu under the model tab & select llama-2-7b-function-calling.Q3_K_M.gguf. Then click "Load" to load the model.
  5. Run cmdh configure and choose the 'text-generation-webui' option.

cmdh will automatically send the prompts to whichever model is loaded by text-generation-webui.

HuggingFace model URL: https://huggingface.co/Trelis/Llama-2-7b-chat-hf-function-calling-v2

Roadmap

The issue tracker is mostly feature requests I have put in so I don't forget them. If you have any bug reports or good ideas, please include them in the tracker.

Issues

If you run into any issues installing or running cmdh, please open a ticket in the project tracker. Include a detailed bug report with stacktraces and inputs and the mode of operation (OpenAI or ollama).

About

Create Linux commands from natural language, in the shell.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published