-
Notifications
You must be signed in to change notification settings - Fork 2k
/
googlevertexai-types.ts
103 lines (85 loc) 路 2.99 KB
/
googlevertexai-types.ts
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
import type { BaseLLMParams } from "@langchain/core/language_models/llms";
export interface GoogleConnectionParams<AuthOptions> {
authOptions?: AuthOptions;
}
export interface GoogleVertexAIConnectionParams<AuthOptions>
extends GoogleConnectionParams<AuthOptions> {
/** Hostname for the API call */
endpoint?: string;
/** Region where the LLM is stored */
location?: string;
/** The version of the API functions. Part of the path. */
apiVersion?: string;
/**
* If you are planning to connect to a model that lives under a custom endpoint
* provide the "customModelURL" which will override the automatic URL building
*
* This is necessary in cases when you want to point to a fine-tuned model or
* a model that has been hidden under VertexAI Endpoints.
*
* In those cases, specifying the `GoogleVertexAIModelParams.model` param
* will not be necessary and will be ignored.
*
* @see GoogleVertexAILLMConnection.buildUrl
* */
customModelURL?: string;
}
export interface GoogleVertexAIModelParams {
/** Model to use */
model?: string;
/** Sampling temperature to use */
temperature?: number;
/**
* Maximum number of tokens to generate in the completion.
*/
maxOutputTokens?: number;
/**
* Top-p changes how the model selects tokens for output.
*
* Tokens are selected from most probable to least until the sum
* of their probabilities equals the top-p value.
*
* For example, if tokens A, B, and C have a probability of
* .3, .2, and .1 and the top-p value is .5, then the model will
* select either A or B as the next token (using temperature).
*/
topP?: number;
/**
* Top-k changes how the model selects tokens for output.
*
* A top-k of 1 means the selected token is the most probable among
* all tokens in the model鈥檚 vocabulary (also called greedy decoding),
* while a top-k of 3 means that the next token is selected from
* among the 3 most probable tokens (using temperature).
*/
topK?: number;
}
export interface GoogleVertexAIBaseLLMInput<AuthOptions>
extends BaseLLMParams,
GoogleVertexAIConnectionParams<AuthOptions>,
GoogleVertexAIModelParams {}
export interface GoogleResponse {
// eslint-disable-next-line @typescript-eslint/no-explicit-any
data: any;
}
export interface GoogleVertexAIBasePrediction {
// eslint-disable-next-line @typescript-eslint/no-explicit-any
safetyAttributes?: any;
}
export interface GoogleVertexAILLMPredictions<
PredictionType extends GoogleVertexAIBasePrediction
> {
predictions: PredictionType[];
}
export type GoogleAbstractedClientOpsMethod = "GET" | "POST";
export type GoogleAbstractedClientOpsResponseType = "json" | "stream";
export type GoogleAbstractedClientOps = {
url?: string;
method?: GoogleAbstractedClientOpsMethod;
data?: unknown;
responseType?: GoogleAbstractedClientOpsResponseType;
};
export interface GoogleAbstractedClient {
request: (opts: GoogleAbstractedClientOps) => unknown;
getProjectId: () => Promise<string>;
}