Skip to content
@InferenceNexus

InferenceNexus

Popular repositories Loading

  1. text-generation-inference text-generation-inference Public

    Forked from huggingface/text-generation-inference

    Large Language Model Text Generation Inference

    Python 1

  2. T-MAC T-MAC Public

    Forked from microsoft/T-MAC

    Low-bit LLM inference on CPU with lookup table

    C++ 1

  3. ipex-llm ipex-llm Public

    Forked from intel/ipex-llm

    Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc,…

    Python 1

  4. litellm litellm Public

    Forked from BerriAI/litellm

    Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]

    Python 1

  5. litgpt litgpt Public

    Forked from Lightning-AI/litgpt

    20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.

    Python 1

  6. inference-benchmarker inference-benchmarker Public

    Forked from huggingface/inference-benchmarker

    Inference server benchmarking tool

    Rust 1

Repositories

Showing 10 of 52 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…