MetalChat is a Metal-accelerated C++ framework and command line interpreter for inference of Meta Llama models.
Important
The library API and CLI are under active development, therefore they may change without any deprecation notice. See issues tab for the list of known issues or missing features.
See the getting started guide for using MetalChat as a library and command line guide for using MetalChat to interact with LLM model from the command line.
The framework and command line utility could be installed using Homebrew package manager in a following way:
brew tap ybubnov/metalchat https://github.com/ybubnov/metalchat
brew install --HEAD metalchatAlternatively you could build a Conan package locally using dependencies downloaded from the Conan registry. After that, you could use the MetalChat framework just like any other Conan dependency.
git clone https://github.com/ybubnov/metalchat
cd metalchat
conan build \
--build=missing \
--settings build_type=Release \
--conf tools.build.skip_test=True \
--options '&:build_executable'=False \
--options '&:use_system_libs'=False
conan export-pkgIf you are using CMake as a build system, you could link the framework using an automatically exported target:
find_package(metalchat CONFIG REQUIRED)
target_link_libraries(build_target PRIVATE MetalChat::MetalChat)The MetalChat is distributed under GPLv3 license. See the LICENSE file for full license text.