diff --git a/docs/supported_models.md b/docs/supported_models.md index 5ed4517d261..e2b6eb48744 100644 --- a/docs/supported_models.md +++ b/docs/supported_models.md @@ -41,6 +41,7 @@ These models accept text input. |⭐QWEN2.5|BF16/WINT8/FP8|Qwen/qwen2.5-72B;
Qwen/qwen2.5-32B;
Qwen/qwen2.5-14B;
Qwen/qwen2.5-7B;
Qwen/qwen2.5-3B;
Qwen/qwen2.5-1.5B;
Qwen/qwen2.5-0.5B, etc.| |⭐QWEN2|BF16/WINT8/FP8|Qwen/Qwen/qwen2-72B;
Qwen/Qwen/qwen2-7B;
Qwen/qwen2-1.5B;
Qwen/qwen2-0.5B;
Qwen/QwQ-32, etc.| |⭐DEEPSEEK|BF16/WINT4|unsloth/DeepSeek-V3.1-BF16;
unsloth/DeepSeek-V3-0324-BF16;
unsloth/DeepSeek-R1-BF16, etc.| +|⭐GPT-OSS|BF16/WINT8|unsloth/gpt-oss-20b-BF16, etc.| ## Multimodal Language Models @@ -49,6 +50,7 @@ These models accept multi-modal inputs (e.g., images and text). |Models|DataType|Example HF Model| |-|-|-| | ERNIE-VL |BF16/WINT4/WINT8| baidu/ERNIE-4.5-VL-424B-A47B-Paddle
 [quick start](./get_started/ernie-4.5-vl.md)   [best practice](./best_practices/ERNIE-4.5-VL-424B-A47B-Paddle.md) ;
baidu/ERNIE-4.5-VL-28B-A3B-Paddle
 [quick start](./get_started/quick_start_vl.md)   [best practice](./best_practices/ERNIE-4.5-VL-28B-A3B-Paddle.md) ;| +| PaddleOCR-VL |BF16/WINT4/WINT8| PaddlePaddle/PaddleOCR-VL
  [best practice](./best_practices/PaddleOCR-VL-0.9B.md) ;| | QWEN-VL |BF16/WINT4/FP8| Qwen/Qwen2.5-VL-72B-Instruct;
Qwen/Qwen2.5-VL-32B-Instruct;
Qwen/Qwen2.5-VL-7B-Instruct;
Qwen/Qwen2.5-VL-3B-Instruct| More models are being supported. You can submit requests for new model support via [Github Issues](https://github.com/PaddlePaddle/FastDeploy/issues). diff --git a/docs/zh/supported_models.md b/docs/zh/supported_models.md index 5caa83ab8aa..dc42414e13a 100644 --- a/docs/zh/supported_models.md +++ b/docs/zh/supported_models.md @@ -39,6 +39,7 @@ python -m fastdeploy.entrypoints.openai.api_server \ |⭐QWEN2.5|BF16/WINT8/FP8|Qwen/qwen2.5-72B;
Qwen/qwen2.5-32B;
Qwen/qwen2.5-14B;
Qwen/qwen2.5-7B;
Qwen/qwen2.5-3B;
Qwen/qwen2.5-1.5B;
Qwen/qwen2.5-0.5B, etc.| |⭐QWEN2|BF16/WINT8/FP8|Qwen/Qwen/qwen2-72B;
Qwen/Qwen/qwen2-7B;
Qwen/qwen2-1.5B;
Qwen/qwen2-0.5B;
Qwen/QwQ-32, etc.| |⭐DEEPSEEK|BF16/WINT4|unsloth/DeepSeek-V3.1-BF16;
unsloth/DeepSeek-V3-0324-BF16;
unsloth/DeepSeek-R1-BF16, etc.| +|⭐GPT-OSS|BF16/WINT8|unsloth/gpt-oss-20b-BF16, etc.| ## 多模态语言模型列表 @@ -47,6 +48,7 @@ python -m fastdeploy.entrypoints.openai.api_server \ |模型|DataType|模型案例| |-|-|-| | ERNIE-VL |BF16/WINT4/WINT8| baidu/ERNIE-4.5-VL-424B-A47B-Paddle
 [快速部署](./get_started/ernie-4.5-vl.md)   [最佳实践](./best_practices/ERNIE-4.5-VL-424B-A47B-Paddle.md) ;
baidu/ERNIE-4.5-VL-28B-A3B-Paddle
 [快速部署](./get_started/quick_start_vl.md)   [最佳实践](./best_practices/ERNIE-4.5-VL-28B-A3B-Paddle.md) ;| +| PaddleOCR-VL |BF16/WINT4/WINT8| PaddlePaddle/PaddleOCR-VL
  [最佳实践](./best_practices/PaddleOCR-VL-0.9B.md) ;| | QWEN-VL |BF16/WINT4/FP8| Qwen/Qwen2.5-VL-72B-Instruct;
Qwen/Qwen2.5-VL-32B-Instruct;
Qwen/Qwen2.5-VL-7B-Instruct;
Qwen/Qwen2.5-VL-3B-Instruct| 更多模型同步支持中,你可以通过[Github Issues](https://github.com/PaddlePaddle/FastDeploy/issues)向我们提交新模型的支持需求。