Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions pages/managed-inference/faq.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -60,6 +60,10 @@ You can select the Instance type based on your model’s computational needs and
Billing is based on the Instance type and usage duration. Unlike [Generative APIs](/generative-apis/quickstart/), which are billed per token, Managed Inference provides predictable costs based on the allocated infrastructure.
Pricing details can be found on the [Scaleway pricing page](https://www.scaleway.com/en/pricing/model-as-a-service/#managed-inference).

## Can I pause Managed Inference billing when the instance is not in use ?
When a Managed Inference deployment is running, corresponding resources are provisioned and thus billed. Resources can therefore not be paused.
However, you can still optimize your Managed Inference deployment to fit within specific time ranges (such as during working hours). To do so, you can automate deployment creation and deletion using the [Managed Inference API](https://www.scaleway.com/en/developers/api/inference/), [Terraform](https://registry.terraform.io/providers/scaleway/scaleway/latest/docs/resources/inference_deployment) or [Scaleway SDKs](https://www.scaleway.com/en/docs/scaleway-sdk/). These actions can be programmed using [Serverless Jobs](/serverless-jobs/) to be automatically carried out periodically.

## Can I run inference on private models?
Yes, Managed Inference allows you to deploy private models with access control settings. You can restrict access to specific users, teams, or networks.

Expand Down