Skip to content

Dieshen/ocelotl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

201 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Ocelotl

Rust-first LLM inference runtime.

Ocelotl is an early-stage workspace for a local LLM runtime with explicit model, loader, tokenizer, kernel, runtime, and serving boundaries. The first milestone is a narrow, correct single-process runtime before adding broad model coverage or high-scale serving features.

Start Here

New contributors should start with docs/start-here.md.

Core orientation docs:

Crates

  • ocelotl-core: shared types, errors, model metadata, and device contracts.
  • ocelotl-loader: model artifact loading and validation.
  • ocelotl-tokenizer: tokenizer and chat-template boundary.
  • ocelotl-kernels: portable kernel dispatch boundary.
  • ocelotl-models: model-family implementations.
  • ocelotl-runtime: request lifecycle, KV cache, scheduling, and generation.
  • ocelotl-server: API/server integration layer.
  • ocelotl: root crate and CLI entrypoint.

Validation

cargo fmt --all
cargo check --workspace
cargo test --workspace

Current Status

This is a project skeleton. Public APIs are intentionally small while the runtime shape is established.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors