From caaf6524b5b07113c0dea1c42a8bf8d06ee3ac88 Mon Sep 17 00:00:00 2001 From: mudler <2420543+mudler@users.noreply.github.com> Date: Tue, 28 Apr 2026 21:18:42 +0000 Subject: [PATCH] feat(swagger): update swagger Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> --- swagger/docs.go | 44 +++++++++++++++++++------------------------- swagger/swagger.json | 44 +++++++++++++++++++------------------------- swagger/swagger.yaml | 34 ++++++++++++++-------------------- 3 files changed, 52 insertions(+), 70 deletions(-) diff --git a/swagger/docs.go b/swagger/docs.go index f1f3f8ae9d41..b5e03ada7f33 100644 --- a/swagger/docs.go +++ b/swagger/docs.go @@ -842,7 +842,7 @@ const docTemplate = `{ "in": "body", "required": true, "schema": { - "$ref": "#/definitions/localai.vramEstimateRequest" + "$ref": "#/definitions/modeladmin.VRAMRequest" } } ], @@ -850,7 +850,7 @@ const docTemplate = `{ "200": { "description": "VRAM estimate", "schema": { - "$ref": "#/definitions/localai.vramEstimateResponse" + "$ref": "#/definitions/modeladmin.VRAMResponse" } } } @@ -3044,36 +3044,45 @@ const docTemplate = `{ } } }, - "localai.vramEstimateRequest": { + "model.BackendLogLine": { + "type": "object", + "properties": { + "stream": { + "description": "\"stdout\" or \"stderr\"", + "type": "string" + }, + "text": { + "type": "string" + }, + "timestamp": { + "type": "string" + } + } + }, + "modeladmin.VRAMRequest": { "type": "object", "properties": { "context_size": { - "description": "context length to estimate for (default 8192)", "type": "integer" }, "gpu_layers": { - "description": "number of layers to offload to GPU (0 = all)", "type": "integer" }, "kv_quant_bits": { - "description": "KV cache quantization bits (0 = fp16)", "type": "integer" }, "model": { - "description": "model name (must be installed)", "type": "string" } } }, - "localai.vramEstimateResponse": { + "modeladmin.VRAMResponse": { "type": "object", "properties": { "context_note": { - "description": "note when context_size was defaulted", "type": "string" }, "model_max_context": { - "description": "model's trained maximum context length", "type": "integer" }, "sizeBytes": { @@ -3094,21 +3103,6 @@ const docTemplate = `{ } } }, - "model.BackendLogLine": { - "type": "object", - "properties": { - "stream": { - "description": "\"stdout\" or \"stderr\"", - "type": "string" - }, - "text": { - "type": "string" - }, - "timestamp": { - "type": "string" - } - } - }, "proto.MemoryUsageData": { "type": "object", "properties": { diff --git a/swagger/swagger.json b/swagger/swagger.json index fd2f2fd1a372..f118aa13c360 100644 --- a/swagger/swagger.json +++ b/swagger/swagger.json @@ -839,7 +839,7 @@ "in": "body", "required": true, "schema": { - "$ref": "#/definitions/localai.vramEstimateRequest" + "$ref": "#/definitions/modeladmin.VRAMRequest" } } ], @@ -847,7 +847,7 @@ "200": { "description": "VRAM estimate", "schema": { - "$ref": "#/definitions/localai.vramEstimateResponse" + "$ref": "#/definitions/modeladmin.VRAMResponse" } } } @@ -3041,36 +3041,45 @@ } } }, - "localai.vramEstimateRequest": { + "model.BackendLogLine": { + "type": "object", + "properties": { + "stream": { + "description": "\"stdout\" or \"stderr\"", + "type": "string" + }, + "text": { + "type": "string" + }, + "timestamp": { + "type": "string" + } + } + }, + "modeladmin.VRAMRequest": { "type": "object", "properties": { "context_size": { - "description": "context length to estimate for (default 8192)", "type": "integer" }, "gpu_layers": { - "description": "number of layers to offload to GPU (0 = all)", "type": "integer" }, "kv_quant_bits": { - "description": "KV cache quantization bits (0 = fp16)", "type": "integer" }, "model": { - "description": "model name (must be installed)", "type": "string" } } }, - "localai.vramEstimateResponse": { + "modeladmin.VRAMResponse": { "type": "object", "properties": { "context_note": { - "description": "note when context_size was defaulted", "type": "string" }, "model_max_context": { - "description": "model's trained maximum context length", "type": "integer" }, "sizeBytes": { @@ -3091,21 +3100,6 @@ } } }, - "model.BackendLogLine": { - "type": "object", - "properties": { - "stream": { - "description": "\"stdout\" or \"stderr\"", - "type": "string" - }, - "text": { - "type": "string" - }, - "timestamp": { - "type": "string" - } - } - }, "proto.MemoryUsageData": { "type": "object", "properties": { diff --git a/swagger/swagger.yaml b/swagger/swagger.yaml index 93a6f582e55f..3b47dac6d3a3 100644 --- a/swagger/swagger.yaml +++ b/swagger/swagger.yaml @@ -313,28 +313,32 @@ definitions: >= 1. type: integer type: object - localai.vramEstimateRequest: + model.BackendLogLine: + properties: + stream: + description: '"stdout" or "stderr"' + type: string + text: + type: string + timestamp: + type: string + type: object + modeladmin.VRAMRequest: properties: context_size: - description: context length to estimate for (default 8192) type: integer gpu_layers: - description: number of layers to offload to GPU (0 = all) type: integer kv_quant_bits: - description: KV cache quantization bits (0 = fp16) type: integer model: - description: model name (must be installed) type: string type: object - localai.vramEstimateResponse: + modeladmin.VRAMResponse: properties: context_note: - description: note when context_size was defaulted type: string model_max_context: - description: model's trained maximum context length type: integer sizeBytes: description: total model weight size in bytes @@ -349,16 +353,6 @@ definitions: description: human-readable VRAM (e.g. "6.1 GB") type: string type: object - model.BackendLogLine: - properties: - stream: - description: '"stdout" or "stderr"' - type: string - text: - type: string - timestamp: - type: string - type: object proto.MemoryUsageData: properties: breakdown: @@ -2710,14 +2704,14 @@ paths: name: request required: true schema: - $ref: '#/definitions/localai.vramEstimateRequest' + $ref: '#/definitions/modeladmin.VRAMRequest' produces: - application/json responses: "200": description: VRAM estimate schema: - $ref: '#/definitions/localai.vramEstimateResponse' + $ref: '#/definitions/modeladmin.VRAMResponse' summary: Estimate VRAM usage for a model tags: - config