Description
Bicep version
0.18.4
Describe the bug
I am fairly confident this is an issue with the native provider. I am attempting to deploy a Cognitive Services account w/ gpt-35-turbo deployment.
The portal shows the following:
{
"type": "Microsoft.CognitiveServices/accounts/deployments",
"apiVersion": "2022-12-01",
"name": "[concat(parameters('accounts_pdfgptdemo_dev_eus_name'), '/gpt-35-turbo')]",
"dependsOn": [
"[resourceId('Microsoft.CognitiveServices/accounts', parameters('accounts_pdfgptdemo_dev_eus_name'))]"
],
"properties": {
"model": {
"format": "OpenAI",
"name": "gpt-35-turbo",
"version": "0301"
},
"scaleSettings": {
"scaleType": "Standard",
"capacity": 120
},
"raiPolicyName": "Microsoft.Default"
}
},
My bicep first attempt was:
resource gpt4Deployment 'Microsoft.CognitiveServices/accounts/deployments@2022-12-01'={
parent: openAI
name: 'gpt-35-turbo'
properties: {
model: {
format: 'OpenAI'
name: 'gpt-35-turbo'
version: '0301'
}
scaleSettings: {
scaleType: 'Standard'
capacity: 120
}
raiPolicyName:'Microsoft.Default'
}
}
I got the
See inner errors for details.\",\r\n \"details\": [\r\n {\r\n \"code\": \"InvalidCapacity\",\r\n \"message\": \"The capacity should be null for standard deployment.
I switched my to capacity: null
and now get "The specified capacity '120' of account deployment is bigger than available capacity '119' for UsageName 'Tokens Per Minute (thousands) - GPT-35-Turbo'
To Reproduce
I have my IaC here: https://github.com/JFolberth/PDFgpt/tree/gpt3 Python code isn't 100% compatible with GPT 35 though IaC will reflect this issue. When running the deployment will fail and still create the Open AI Model
Additional context
Add any other context about the problem here.
Metadata
Metadata
Assignees
Type
Projects
Status