From 68d7a9a7a8024b636fe71dc2eaec1b4dd35257c1 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Mon, 15 Sep 2025 23:24:10 +0200 Subject: [PATCH 01/11] fix: add max pages to llm args --- args/llm.go | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/args/llm.go b/args/llm.go index 1816094..e483aeb 100644 --- a/args/llm.go +++ b/args/llm.go @@ -12,6 +12,7 @@ var ( ErrLLMDatasetIdRequired = errors.New("dataset id is required") ErrLLMPromptRequired = errors.New("prompt is required") ErrLLMMaxTokensNegative = errors.New("max tokens must be non-negative") + ErrLLMMaxPagesNegative = errors.New("max pages must be non-negative") ) const ( @@ -19,6 +20,7 @@ const ( LLMDefaultTemperature = "0.1" LLMDefaultMultipleColumns = false LLMDefaultModel = "gemini-1.5-flash-8b" + LLMDefaultMaxPages = 1 ) type LLMProcessorArguments struct { @@ -26,6 +28,7 @@ type LLMProcessorArguments struct { Prompt string `json:"prompt"` MaxTokens int `json:"max_tokens"` Temperature string `json:"temperature"` + MaxPages int `json:"max_pages"` } // UnmarshalJSON implements custom JSON unmarshaling with validation @@ -54,6 +57,9 @@ func (l *LLMProcessorArguments) setDefaultValues() { if l.Temperature == "" { l.Temperature = LLMDefaultTemperature } + if l.MaxPages == 0 { + l.MaxPages = LLMDefaultMaxPages + } } func (l *LLMProcessorArguments) Validate() error { @@ -66,6 +72,9 @@ func (l *LLMProcessorArguments) Validate() error { if l.MaxTokens < 0 { return fmt.Errorf("%w: got %v", ErrLLMMaxTokensNegative, l.MaxTokens) } + if l.MaxPages < 1 { + return fmt.Errorf("%w: got %v", ErrLLMMaxPagesNegative, l.MaxPages) + } return nil } From 8ad23d0c9eb60302b44f5d9728099feca7c76a2d Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Mon, 15 Sep 2025 23:26:35 +0200 Subject: [PATCH 02/11] chore: fix test --- args/llm_test.go | 1 + 1 file changed, 1 insertion(+) diff --git a/args/llm_test.go b/args/llm_test.go index 3884ebf..1b620a0 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -62,6 +62,7 @@ var _ = Describe("LLMProcessorArguments", func() { Prompt: "p", MaxTokens: 10, Temperature: "0.2", + MaxPages: 1, } err := llmArgs.Validate() Expect(err).ToNot(HaveOccurred()) From 95ae253e50f9af8cbcea2664beb42996e1e19277 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 00:16:56 +0200 Subject: [PATCH 03/11] fix: rename to items --- args/llm.go | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/args/llm.go b/args/llm.go index e483aeb..3997110 100644 --- a/args/llm.go +++ b/args/llm.go @@ -28,7 +28,7 @@ type LLMProcessorArguments struct { Prompt string `json:"prompt"` MaxTokens int `json:"max_tokens"` Temperature string `json:"temperature"` - MaxPages int `json:"max_pages"` + Items int `json:"items"` } // UnmarshalJSON implements custom JSON unmarshaling with validation @@ -57,8 +57,8 @@ func (l *LLMProcessorArguments) setDefaultValues() { if l.Temperature == "" { l.Temperature = LLMDefaultTemperature } - if l.MaxPages == 0 { - l.MaxPages = LLMDefaultMaxPages + if l.Items == 0 { + l.Items = LLMDefaultMaxPages } } @@ -72,8 +72,8 @@ func (l *LLMProcessorArguments) Validate() error { if l.MaxTokens < 0 { return fmt.Errorf("%w: got %v", ErrLLMMaxTokensNegative, l.MaxTokens) } - if l.MaxPages < 1 { - return fmt.Errorf("%w: got %v", ErrLLMMaxPagesNegative, l.MaxPages) + if l.Items < 1 { + return fmt.Errorf("%w: got %v", ErrLLMMaxPagesNegative, l.Items) } return nil } From 1ea688bcd6df67f2da47826789d3975964f88a28 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 00:17:07 +0200 Subject: [PATCH 04/11] chore: fix test --- args/llm_test.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/args/llm_test.go b/args/llm_test.go index 1b620a0..4f04020 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -62,7 +62,7 @@ var _ = Describe("LLMProcessorArguments", func() { Prompt: "p", MaxTokens: 10, Temperature: "0.2", - MaxPages: 1, + Items: 1, } err := llmArgs.Validate() Expect(err).ToNot(HaveOccurred()) From 318c9572cc6101fc32a59230350c021ba8a5adfe Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 00:46:24 +0200 Subject: [PATCH 05/11] fix: rename vars --- args/llm.go | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/args/llm.go b/args/llm.go index 3997110..c4da6c5 100644 --- a/args/llm.go +++ b/args/llm.go @@ -12,7 +12,7 @@ var ( ErrLLMDatasetIdRequired = errors.New("dataset id is required") ErrLLMPromptRequired = errors.New("prompt is required") ErrLLMMaxTokensNegative = errors.New("max tokens must be non-negative") - ErrLLMMaxPagesNegative = errors.New("max pages must be non-negative") + ErrLLMItemsNegative = errors.New("items must be non-negative") ) const ( @@ -20,7 +20,7 @@ const ( LLMDefaultTemperature = "0.1" LLMDefaultMultipleColumns = false LLMDefaultModel = "gemini-1.5-flash-8b" - LLMDefaultMaxPages = 1 + LLMDefaultItems = 1 ) type LLMProcessorArguments struct { @@ -58,7 +58,7 @@ func (l *LLMProcessorArguments) setDefaultValues() { l.Temperature = LLMDefaultTemperature } if l.Items == 0 { - l.Items = LLMDefaultMaxPages + l.Items = LLMDefaultItems } } @@ -73,7 +73,7 @@ func (l *LLMProcessorArguments) Validate() error { return fmt.Errorf("%w: got %v", ErrLLMMaxTokensNegative, l.MaxTokens) } if l.Items < 1 { - return fmt.Errorf("%w: got %v", ErrLLMMaxPagesNegative, l.Items) + return fmt.Errorf("%w: got %v", ErrLLMItemsNegative, l.Items) } return nil } From 41bedf3324c0e129447d1cbc356366a5fd428312 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 18:48:20 +0200 Subject: [PATCH 06/11] chore: relevant llm args to uint instead of int --- args/llm.go | 18 +++++------------- args/llm_test.go | 12 ------------ types/llm.go | 2 +- 3 files changed, 6 insertions(+), 26 deletions(-) diff --git a/args/llm.go b/args/llm.go index c4da6c5..da89cf8 100644 --- a/args/llm.go +++ b/args/llm.go @@ -11,8 +11,6 @@ import ( var ( ErrLLMDatasetIdRequired = errors.New("dataset id is required") ErrLLMPromptRequired = errors.New("prompt is required") - ErrLLMMaxTokensNegative = errors.New("max tokens must be non-negative") - ErrLLMItemsNegative = errors.New("items must be non-negative") ) const ( @@ -26,9 +24,9 @@ const ( type LLMProcessorArguments struct { DatasetId string `json:"dataset_id"` Prompt string `json:"prompt"` - MaxTokens int `json:"max_tokens"` + MaxTokens uint `json:"max_tokens"` Temperature string `json:"temperature"` - Items int `json:"items"` + Items uint `json:"items"` } // UnmarshalJSON implements custom JSON unmarshaling with validation @@ -51,12 +49,12 @@ func (l *LLMProcessorArguments) UnmarshalJSON(data []byte) error { } func (l *LLMProcessorArguments) setDefaultValues() { - if l.MaxTokens == 0 { - l.MaxTokens = LLMDefaultMaxTokens - } if l.Temperature == "" { l.Temperature = LLMDefaultTemperature } + if l.MaxTokens == 0 { + l.MaxTokens = LLMDefaultMaxTokens + } if l.Items == 0 { l.Items = LLMDefaultItems } @@ -69,12 +67,6 @@ func (l *LLMProcessorArguments) Validate() error { if l.Prompt == "" { return ErrLLMPromptRequired } - if l.MaxTokens < 0 { - return fmt.Errorf("%w: got %v", ErrLLMMaxTokensNegative, l.MaxTokens) - } - if l.Items < 1 { - return fmt.Errorf("%w: got %v", ErrLLMItemsNegative, l.Items) - } return nil } diff --git a/args/llm_test.go b/args/llm_test.go index 4f04020..6370528 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -87,18 +87,6 @@ var _ = Describe("LLMProcessorArguments", func() { err := llmArgs.Validate() Expect(errors.Is(err, args.ErrLLMPromptRequired)).To(BeTrue()) }) - - It("should fail when max tokens is negative", func() { - llmArgs := &args.LLMProcessorArguments{ - DatasetId: "ds1", - Prompt: "p", - MaxTokens: -1, - Temperature: "0.2", - } - err := llmArgs.Validate() - Expect(errors.Is(err, args.ErrLLMMaxTokensNegative)).To(BeTrue()) - Expect(err.Error()).To(ContainSubstring("got -1")) - }) }) Describe("ToLLMProcessorRequest", func() { diff --git a/types/llm.go b/types/llm.go index fb67693..d8b76a4 100644 --- a/types/llm.go +++ b/types/llm.go @@ -7,7 +7,7 @@ type LLMProcessorRequest struct { MultipleColumns bool `json:"multipleColumns"` Prompt string `json:"prompt"` // example: summarize the content of this webpage: ${markdown} Temperature string `json:"temperature"` - MaxTokens int `json:"maxTokens"` + MaxTokens uint `json:"maxTokens"` } type LLMProcessorResult struct { From dbfee0472339dff0811dfee24fb3c1d4ac3fa572 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 19:00:59 +0200 Subject: [PATCH 07/11] chore: temperature as a float --- args/llm.go | 25 +++++++++++++------------ args/llm_test.go | 12 ++++++------ 2 files changed, 19 insertions(+), 18 deletions(-) diff --git a/args/llm.go b/args/llm.go index da89cf8..2f5c8c6 100644 --- a/args/llm.go +++ b/args/llm.go @@ -4,6 +4,7 @@ import ( "encoding/json" "errors" "fmt" + "strconv" teetypes "github.com/masa-finance/tee-types/types" ) @@ -14,19 +15,19 @@ var ( ) const ( - LLMDefaultMaxTokens = 300 - LLMDefaultTemperature = "0.1" - LLMDefaultMultipleColumns = false - LLMDefaultModel = "gemini-1.5-flash-8b" - LLMDefaultItems = 1 + LLMDefaultMaxTokens uint = 300 + LLMDefaultTemperature float64 = 0.1 + LLMDefaultMultipleColumns bool = false + LLMDefaultModel string = "gemini-1.5-flash-8b" + LLMDefaultItems uint = 1 ) type LLMProcessorArguments struct { - DatasetId string `json:"dataset_id"` - Prompt string `json:"prompt"` - MaxTokens uint `json:"max_tokens"` - Temperature string `json:"temperature"` - Items uint `json:"items"` + DatasetId string `json:"dataset_id"` + Prompt string `json:"prompt"` + MaxTokens uint `json:"max_tokens"` + Temperature float64 `json:"temperature"` + Items uint `json:"items"` } // UnmarshalJSON implements custom JSON unmarshaling with validation @@ -49,7 +50,7 @@ func (l *LLMProcessorArguments) UnmarshalJSON(data []byte) error { } func (l *LLMProcessorArguments) setDefaultValues() { - if l.Temperature == "" { + if l.Temperature == 0 { l.Temperature = LLMDefaultTemperature } if l.MaxTokens == 0 { @@ -75,7 +76,7 @@ func (l LLMProcessorArguments) ToLLMProcessorRequest() teetypes.LLMProcessorRequ InputDatasetId: l.DatasetId, Prompt: l.Prompt, MaxTokens: l.MaxTokens, - Temperature: l.Temperature, + Temperature: strconv.FormatFloat(l.Temperature, 'f', -1, 64), MultipleColumns: LLMDefaultMultipleColumns, // overrides default in actor API Model: LLMDefaultModel, // overrides default in actor API } diff --git a/args/llm_test.go b/args/llm_test.go index 6370528..aa49935 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -30,7 +30,7 @@ var _ = Describe("LLMProcessorArguments", func() { DatasetId: "ds1", Prompt: "summarize: ${markdown}", MaxTokens: 123, - Temperature: "0.7", + Temperature: 0.7, } jsonData, err := json.Marshal(llmArgs) Expect(err).ToNot(HaveOccurred()) @@ -61,7 +61,7 @@ var _ = Describe("LLMProcessorArguments", func() { DatasetId: "ds1", Prompt: "p", MaxTokens: 10, - Temperature: "0.2", + Temperature: 0.2, Items: 1, } err := llmArgs.Validate() @@ -72,7 +72,7 @@ var _ = Describe("LLMProcessorArguments", func() { llmArgs := &args.LLMProcessorArguments{ Prompt: "p", MaxTokens: 10, - Temperature: "0.2", + Temperature: 0.2, } err := llmArgs.Validate() Expect(errors.Is(err, args.ErrLLMDatasetIdRequired)).To(BeTrue()) @@ -82,7 +82,7 @@ var _ = Describe("LLMProcessorArguments", func() { llmArgs := &args.LLMProcessorArguments{ DatasetId: "ds1", MaxTokens: 10, - Temperature: "0.2", + Temperature: 0.2, } err := llmArgs.Validate() Expect(errors.Is(err, args.ErrLLMPromptRequired)).To(BeTrue()) @@ -95,7 +95,7 @@ var _ = Describe("LLMProcessorArguments", func() { DatasetId: "ds1", Prompt: "p", MaxTokens: 0, // default applied in To* - Temperature: "", + Temperature: 0, } req := llmArgs.ToLLMProcessorRequest() Expect(req.InputDatasetId).To(Equal("ds1")) @@ -111,7 +111,7 @@ var _ = Describe("LLMProcessorArguments", func() { DatasetId: "ds1", Prompt: "p", MaxTokens: 42, - Temperature: "0.7", + Temperature: 0.7, } req := llmArgs.ToLLMProcessorRequest() Expect(req.InputDatasetId).To(Equal("ds1")) From fb309279721e2cba0b3a82695cf68fa3f13444f5 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 19:11:42 +0200 Subject: [PATCH 08/11] chore: fix test --- args/llm_test.go | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/args/llm_test.go b/args/llm_test.go index aa49935..a0461c3 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -22,7 +22,7 @@ var _ = Describe("LLMProcessorArguments", func() { err = json.Unmarshal([]byte(jsonData), &llmArgs) Expect(err).ToNot(HaveOccurred()) Expect(llmArgs.MaxTokens).To(Equal(300)) - Expect(llmArgs.Temperature).To(Equal("0.1")) + Expect(llmArgs.Temperature).To(Equal(0.1)) }) It("should override default values", func() { @@ -37,7 +37,7 @@ var _ = Describe("LLMProcessorArguments", func() { err = json.Unmarshal([]byte(jsonData), &llmArgs) Expect(err).ToNot(HaveOccurred()) Expect(llmArgs.MaxTokens).To(Equal(123)) - Expect(llmArgs.Temperature).To(Equal("0.7")) + Expect(llmArgs.Temperature).To(Equal(0.7)) }) It("should fail unmarshal when dataset_id is missing", func() { @@ -101,7 +101,7 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(req.InputDatasetId).To(Equal("ds1")) Expect(req.Prompt).To(Equal("p")) Expect(req.MaxTokens).To(Equal(0)) - Expect(req.Temperature).To(Equal("")) + Expect(req.Temperature).To(Equal(0)) Expect(req.MultipleColumns).To(BeFalse()) Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) }) @@ -117,7 +117,7 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(req.InputDatasetId).To(Equal("ds1")) Expect(req.Prompt).To(Equal("p")) Expect(req.MaxTokens).To(Equal(42)) - Expect(req.Temperature).To(Equal("0.7")) + Expect(req.Temperature).To(Equal(0.7)) Expect(req.MultipleColumns).To(BeFalse()) Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) }) From 15c8eb35c10d75da4a1ce7790b8d4ac790fe3746 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 19:16:58 +0200 Subject: [PATCH 09/11] chore: fix llm test --- args/llm_test.go | 11 +++++++---- 1 file changed, 7 insertions(+), 4 deletions(-) diff --git a/args/llm_test.go b/args/llm_test.go index a0461c3..4b4be90 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -21,8 +21,9 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(err).ToNot(HaveOccurred()) err = json.Unmarshal([]byte(jsonData), &llmArgs) Expect(err).ToNot(HaveOccurred()) - Expect(llmArgs.MaxTokens).To(Equal(300)) + Expect(llmArgs.MaxTokens).To(Equal(uint(300))) Expect(llmArgs.Temperature).To(Equal(0.1)) + Expect(llmArgs.Items).To(Equal(uint(1))) }) It("should override default values", func() { @@ -31,12 +32,14 @@ var _ = Describe("LLMProcessorArguments", func() { Prompt: "summarize: ${markdown}", MaxTokens: 123, Temperature: 0.7, + Items: 3, } jsonData, err := json.Marshal(llmArgs) Expect(err).ToNot(HaveOccurred()) err = json.Unmarshal([]byte(jsonData), &llmArgs) Expect(err).ToNot(HaveOccurred()) - Expect(llmArgs.MaxTokens).To(Equal(123)) + Expect(llmArgs.MaxTokens).To(Equal(uint(123))) + Expect(llmArgs.Items).To(Equal(uint(3))) Expect(llmArgs.Temperature).To(Equal(0.7)) }) @@ -100,7 +103,7 @@ var _ = Describe("LLMProcessorArguments", func() { req := llmArgs.ToLLMProcessorRequest() Expect(req.InputDatasetId).To(Equal("ds1")) Expect(req.Prompt).To(Equal("p")) - Expect(req.MaxTokens).To(Equal(0)) + Expect(req.MaxTokens).To(Equal(uint(0))) Expect(req.Temperature).To(Equal(0)) Expect(req.MultipleColumns).To(BeFalse()) Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) @@ -116,7 +119,7 @@ var _ = Describe("LLMProcessorArguments", func() { req := llmArgs.ToLLMProcessorRequest() Expect(req.InputDatasetId).To(Equal("ds1")) Expect(req.Prompt).To(Equal("p")) - Expect(req.MaxTokens).To(Equal(42)) + Expect(req.MaxTokens).To(Equal(uint(42))) Expect(req.Temperature).To(Equal(0.7)) Expect(req.MultipleColumns).To(BeFalse()) Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) From 44ff24a6431d203225e65b833dfd07fdcab1e6da Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 19:44:33 +0200 Subject: [PATCH 10/11] fix: llm test --- args/llm_test.go | 4 ++-- types/llm.go | 4 ++-- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/args/llm_test.go b/args/llm_test.go index 4b4be90..b3a9007 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -104,7 +104,7 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(req.InputDatasetId).To(Equal("ds1")) Expect(req.Prompt).To(Equal("p")) Expect(req.MaxTokens).To(Equal(uint(0))) - Expect(req.Temperature).To(Equal(0)) + Expect(req.Temperature).To(Equal("0.1")) Expect(req.MultipleColumns).To(BeFalse()) Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) }) @@ -120,7 +120,7 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(req.InputDatasetId).To(Equal("ds1")) Expect(req.Prompt).To(Equal("p")) Expect(req.MaxTokens).To(Equal(uint(42))) - Expect(req.Temperature).To(Equal(0.7)) + Expect(req.Temperature).To(Equal("0.7")) Expect(req.MultipleColumns).To(BeFalse()) Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) }) diff --git a/types/llm.go b/types/llm.go index d8b76a4..ca99075 100644 --- a/types/llm.go +++ b/types/llm.go @@ -5,8 +5,8 @@ type LLMProcessorRequest struct { LLMProviderApiKey string `json:"llmProviderApiKey"` // encrypted api key by miner Model string `json:"model"` MultipleColumns bool `json:"multipleColumns"` - Prompt string `json:"prompt"` // example: summarize the content of this webpage: ${markdown} - Temperature string `json:"temperature"` + Prompt string `json:"prompt"` // example: summarize the content of this webpage: ${markdown} + Temperature string `json:"temperature"` // the actor expects a string MaxTokens uint `json:"maxTokens"` } From 4466953b9926b6d98c8b78020c39c6bb11de9962 Mon Sep 17 00:00:00 2001 From: grantdfoster Date: Tue, 16 Sep 2025 19:51:39 +0200 Subject: [PATCH 11/11] fix: temperature test --- args/llm_test.go | 22 +++------------------- 1 file changed, 3 insertions(+), 19 deletions(-) diff --git a/args/llm_test.go b/args/llm_test.go index b3a9007..aa35128 100644 --- a/args/llm_test.go +++ b/args/llm_test.go @@ -21,8 +21,8 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(err).ToNot(HaveOccurred()) err = json.Unmarshal([]byte(jsonData), &llmArgs) Expect(err).ToNot(HaveOccurred()) - Expect(llmArgs.MaxTokens).To(Equal(uint(300))) Expect(llmArgs.Temperature).To(Equal(0.1)) + Expect(llmArgs.MaxTokens).To(Equal(uint(300))) Expect(llmArgs.Items).To(Equal(uint(1))) }) @@ -38,9 +38,9 @@ var _ = Describe("LLMProcessorArguments", func() { Expect(err).ToNot(HaveOccurred()) err = json.Unmarshal([]byte(jsonData), &llmArgs) Expect(err).ToNot(HaveOccurred()) + Expect(llmArgs.Temperature).To(Equal(0.7)) Expect(llmArgs.MaxTokens).To(Equal(uint(123))) Expect(llmArgs.Items).To(Equal(uint(3))) - Expect(llmArgs.Temperature).To(Equal(0.7)) }) It("should fail unmarshal when dataset_id is missing", func() { @@ -93,23 +93,7 @@ var _ = Describe("LLMProcessorArguments", func() { }) Describe("ToLLMProcessorRequest", func() { - It("should map fields and defaults correctly", func() { - llmArgs := args.LLMProcessorArguments{ - DatasetId: "ds1", - Prompt: "p", - MaxTokens: 0, // default applied in To* - Temperature: 0, - } - req := llmArgs.ToLLMProcessorRequest() - Expect(req.InputDatasetId).To(Equal("ds1")) - Expect(req.Prompt).To(Equal("p")) - Expect(req.MaxTokens).To(Equal(uint(0))) - Expect(req.Temperature).To(Equal("0.1")) - Expect(req.MultipleColumns).To(BeFalse()) - Expect(req.Model).To(Equal("gemini-1.5-flash-8b")) - }) - - It("should map fields correctly when set", func() { + It("should map request fields to actor request fields", func() { llmArgs := args.LLMProcessorArguments{ DatasetId: "ds1", Prompt: "p",