Skip to content
View greasymonster's full-sized avatar

Block or report greasymonster

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Popular repositories Loading

  1. Chat_with_Datawhale_langchain Chat_with_Datawhale_langchain Public

    Forked from logan-zou/Chat_with_Datawhale_langchain

    Python

  2. PromptCBLUE PromptCBLUE Public

    Forked from michael-wzhu/PromptCBLUE

    PromptCBLUE: a large-scale instruction-tuning dataset for multi-task and few-shot learning in the medical domain in Chinese

    Python

  3. ann-benchmarks ann-benchmarks Public

    Forked from erikbern/ann-benchmarks

    Benchmarks of approximate nearest neighbor libraries in Python

    Python

  4. llama2.c llama2.c Public

    Forked from karpathy/llama2.c

    Inference Llama 2 in one file of pure C

    C

  5. LightLLM LightLLM Public

    Forked from ModelTC/LightLLM

    LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

    Python

  6. fastllm fastllm Public

    Forked from ztxz16/fastllm

    fastllm是后端无依赖的高性能大模型推理库。同时支持张量并行推理稠密模型和混合模式推理MOE模型,任意10G以上显卡即可推理满血DeepSeek。双路9004/9005服务器+单显卡部署DeepSeek满血满精度原版模型,单并发20tps;INT4量化模型单并发30tps,多并发可达60+。

    C++