diff --git a/pages/managed-inference/reference-content/supported-models.mdx b/pages/managed-inference/reference-content/supported-models.mdx index 1ac29d77c9..4cd80f16c8 100644 --- a/pages/managed-inference/reference-content/supported-models.mdx +++ b/pages/managed-inference/reference-content/supported-models.mdx @@ -9,7 +9,7 @@ dates: Scaleway Managed Inference allows you to deploy various AI models, either from: - * [Scaleway model catalog](#scaleway-model-catalog): A curated set of ready-to-deploy models available through the [Scaleway console](https:/console.scaleway.com/inference/deployments/) or the [Managed Inference models API](https:/www.scaleway.com/en/developers/api/inference/#path-models-list-models) + * [Scaleway model catalog](#scaleway-model-catalog): A curated set of ready-to-deploy models available through the [Scaleway console](https://console.scaleway.com/inference/deployments/) or the [Managed Inference models API](https://www.scaleway.com/en/developers/api/inference/#path-models-list-models) * [Custom models](#custom-models): Models that you import, typically from sources like Hugging Face. ## Scaleway model catalog @@ -19,14 +19,14 @@ You can find a complete list of all models available in Scaleway's catalog on th ## Custom models - Custom model support is currently in **beta**. If you encounter issues or limitations, please report them via our [Slack community channel](https:/scaleway-community.slack.com/archives/C01SGLGRLEA) or [customer support](https:/console.scaleway.com/support/tickets/create?for=product&productName=inference). + Custom model support is currently in **beta**. If you encounter issues or limitations, please report them via our [Slack community channel](https://scaleway-community.slack.com/archives/C01SGLGRLEA) or [customer support](https://console.scaleway.com/support/tickets/create?for=product&productName=inference). ### Prerequisites We recommend starting with a variation of a supported model from the Scaleway catalog. - For example, you can deploy a [quantized (4-bit) version of Llama 3.3](https:/huggingface.co/unsloth/Llama-3.3-70B-Instruct-bnb-4bit). + For example, you can deploy a [quantized (4-bit) version of Llama 3.3](https://huggingface.co/unsloth/Llama-3.3-70B-Instruct-bnb-4bit). If deploying a fine-tuned version of Llama 3.3, make sure your file structure matches the example linked above. Examples whose compatibility has been tested are available in [tested models](#known-compatible-models). @@ -37,7 +37,7 @@ To deploy a custom model via Hugging Face, ensure the following: * You must have access to the model using your Hugging Face credentials. * For gated models, request access through your Hugging Face account. - * Credentials are not stored, but we recommend using [read or fine-grained access tokens](https:/huggingface.co/docs/hub/security-tokens). + * Credentials are not stored, but we recommend using [read or fine-grained access tokens](https://huggingface.co/docs/hub/security-tokens). #### Required files @@ -46,7 +46,7 @@ Your model repository must include: * A `config.json` file containig: * An `architectures` array (see [supported architectures](#supported-models-architecture) for the exact list of supported values). * `max_position_embeddings` - * Model weights in the [`.safetensors`](https:/huggingface.co/docs/safetensors/index) format + * Model weights in the [`.safetensors`](https://huggingface.co/docs/safetensors/index) format * A `tokenizer.json` file * If your are fine-tuning an existing model, we recommend you use the same `tokenizer.json` file from the base model. * A chat template included in either: @@ -68,7 +68,7 @@ Your model must be one of the following types: **Security Notice**
- Models using formats that allow arbitrary code execution, such as Python [`pickle`](https:/docs.python.org/3/library/pickle.html), are **not supported**. + Models using formats that allow arbitrary code execution, such as Python [`pickle`](https://docs.python.org/3/library/pickle.html), are **not supported**.
## API support