Magic CLI is a command line utility which uses LLMs to help you use the command line more efficiently, inspired by projects such as Amazon Q (prev. Fig terminal) and GitHub Copilot for CLI.
magic-cli
relies on the orch library for LLM interactions (execution, orchestration, model alignment, etc.).
Read the announcement blog post.
Caution
This project is still in early development.
Expect breaking changes and bugs, and please report any issues you encounter.
Thank you!
- Suggest a command (see section)
- Ask to generate a command to perform a task (see section)
- Semantic search of commands across your shell history
- Use a local or remote LLM (see section)
Note
For more options on how to install Magic CLI, see the releases page for the version you wish to install.
curl -LsSf https://github.com/guywaldman/magic-cli/releases/download/0.0.6/magic-cli-installer.sh | sh
brew install guywaldman/tap/magic-cli
powershell -c "irm https://github.com/guywaldman/magic-cli/releases/download/0.0.6/magic-cli-installer.ps1 | iex"
See the releases page for binaries for your platform.
Add functions to your ~/.bashrc
or ~/.zshrc
to make formatting the prompts in the terminal easier (not requiring quotes).
For example:
function mcs {
model_prompt="$*"
magic-cli suggest "$model_prompt"
}
function mcf {
model_prompt="$*"
magic-cli search "$model_prompt"
}
function mca {
model_prompt="$*"
magic-cli ask "$model_prompt"
}
Supply a prompt to get a suggestion for the command to run.
This is useful in scenarios where you know approximately what you want (and perhaps even the tool) but don't remember the exact arguments or their order.
This is especially useful with CLI tools like ffmpeg
or kubectl
.
magic-cli suggest "Resize test_image.png to 300x300 with ffmpeg"
Usage: magic-cli suggest <PROMPT>
Arguments:
<PROMPT> The prompt to suggest a command for (e.g., "Resize image to 300x300 with ffmpeg")
Search a command across your shell history, and get a list of the top results.
magic-cli search "zellij attach"
Usage: magic-cli search [OPTIONS] <PROMPT>
Arguments:
<PROMPT> The prompt to search for
Important
Word to the wise: If you're using a non-local LLM, be wary about the cost of the embeddings, especially for long shell histories.
Supply a prompt with a task you want the model to perform, and watch it try to suggest a command to run to achieve the goal. It may prompt you along the way to run commands if it needs more context.
magic-cli ask "Set up the dev environment as described in the README"
Usage: magic-cli ask <PROMPT>
Arguments:
<PROMPT> The prompt to ask for (e.g., "Set up the development environment")
Magic CLI supports two LLM providers:
ollama
: Ollama is a local LLM provider. The command expects Ollama to be installed and running on your local machine.openai
: OpenAI is a cloud LLM provider. You configure an API token, and Magic CLI uses it with the OpenAI APIs.
Magic CLI stores configurations in ~/.config/magic_cli/config.json
.
Use magic-cli config
(to see the options, use magic-cli config --help
) to set the configuration options:
Usage: magic-cli config <COMMAND>
Commands:
set Set a value
get Get a value
list List the configurations
reset Reset the configurations to the default values
path Get the path to the configuration file
The currently suppprted configuration options are:
llm
: The LLM to use for generating responses. Supported values: "ollama", "openai"ollama.base_url
: The base URL of the Ollama API (default: "http://localhost:11434")ollama.embedding_model
: The model to use for generating embeddings (default: "nomic-embed-text:latest")ollama.model
: The model to use for generating responses (default: "codestral:latest")openai.api_key
(secret): The API key for the OpenAI APIopenai.embedding_model
: The model to use for generating embeddings (default: "text-embedding-ada-002")openai.model
: The model to use for generating responses (default: "gpt-4o")suggest.add_to_history
: Whether to add the suggested command to the shell history (default: false)suggest.mode
: The mode to use for suggesting commands. Supported values: "clipboard" (copying command to clipboard), "unsafe-execution" (executing in the current shell session) (default: "unsafe-execution")Note:
unsafe-execution
is named as such to make it clear that the CLI executes the command in the current shell session. Please be extremely careful when using this mode - Magic CLI is not responsible for the execution of commands suggested.
-
Windows support (PowerShell is supported, but Windows has not been tested properly)
-
Support for more LLM providers (e.g., Anthropic)
-
Improve local embedding index (currently stored naively as a JSON, looked into SQLLite with vector extensions)
-
More test coverage
Security is taken seriously and all vulnerabilities will be handled with utmost care and priority.
In terms of data stored, the sensitive data that is currently handled by Magic CLI is:
- OpenAI API key, which is stored in the configuration within the user home directory (
~/.config/magic_cli
).There are plans to store this token in the system's secure key storage, but this is not yet implemented.
- Embeddings of the shell history for the
magic-cli search
command, which are stored in the configuration within the user home directory (~/.config/magic_cli
) and are generated using the LLM provider of choice.
Please see SECURITY.md for more information, and instructions on how to report potential vulnerabilities.
Contributions are welcome!
Please see CONTRIBUTING.md for more information.
My only request is that pull requests follow an issue, such that we avoid situations of your hard work not being accepted due to a lack of context or necessity.