Skip to content

llm plugin for Cerebras fast inference API

License

Notifications You must be signed in to change notification settings

Olshansk/llm-cerebras

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PyPI Changelog License

llm plugin to prompt Cerebras hosted models.

Install this plugin in the same environment as LLM:

llm install llm-cerebras

You'll need to obtain a Cerebras API key. Once you have it, configure the plugin like this:

llm keys set cerebras

To use the Cerebras models, run:

llm -m llama3.1-8b "Your prompt here"

Or for the 70B model:

llm -m llama3.1-70b "Your prompt here"

The following options are available:

  • temperature: Controls randomness. Defaults to 0.7, range 0-1.5.
  • max_tokens: The maximum number of tokens to generate.
  • top_p: Alternative to temperature for nucleus sampling. Defaults to 1.
  • seed: For deterministic sampling.

Example usage with options:

llm -m llama3.1-8b "Your prompt" -o temperature 0.5 -o max_tokens 100

To set up this plugin locally, first checkout the code. Then create a new virtual environment:

cd llm-cerebras
python3 -m venv venv
source venv/bin/activate

Now install the dependencies and test dependencies:

pip install -e '.[test]'

To run the tests:

pytest

About

llm plugin for Cerebras fast inference API

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%