File tree 7 files changed +32
-8
lines changed
7 files changed +32
-8
lines changed Original file line number Diff line number Diff line change 1
1
configured_endpoints : 16
2
- openapi_spec_url : https://storage.googleapis.com/stainless-sdk-openapi-specs/togetherai%2FTogetherAI-dac0349a130b23065cbd771f383ef7eb7fa99ba35cae0da6be2f55f82d2f6d3a .yml
2
+ openapi_spec_url : https://storage.googleapis.com/stainless-sdk-openapi-specs/togetherai%2FTogetherAI-3950390ec89ef1e5aec351d8e841d511ebc0a3ccec5bbf0d4c73157d1f9d7c1d .yml
Original file line number Diff line number Diff line change @@ -250,6 +250,8 @@ export interface CompletionCreateParamsBase {
250
250
251
251
/**
252
252
* The name of the model to query.
253
+ *
254
+ * [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
253
255
*/
254
256
model :
255
257
| 'Qwen/Qwen2.5-72B-Instruct-Turbo'
Original file line number Diff line number Diff line change @@ -123,8 +123,15 @@ export type CompletionCreateParams = CompletionCreateParamsNonStreaming | Comple
123
123
export interface CompletionCreateParamsBase {
124
124
/**
125
125
* The name of the model to query.
126
+ *
127
+ * [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
126
128
*/
127
- model : string ;
129
+ model :
130
+ | 'meta-llama/Llama-2-70b-hf'
131
+ | 'mistralai/Mistral-7B-v0.1'
132
+ | 'mistralai/Mixtral-8x7B-v0.1'
133
+ | 'Meta-Llama/Llama-Guard-7b'
134
+ | ( string & { } ) ;
128
135
129
136
/**
130
137
* A string providing context for the model to complete.
@@ -186,7 +193,7 @@ export interface CompletionCreateParamsBase {
186
193
* available moderation models found
187
194
* [here](https://docs.together.ai/docs/inference-models#moderation-models).
188
195
*/
189
- safety_model ?: string ;
196
+ safety_model ?: 'Meta-Llama/Llama-Guard-7b' | ( string & { } ) ;
190
197
191
198
/**
192
199
* Seed value for reproducibility.
Original file line number Diff line number Diff line change @@ -39,8 +39,15 @@ export interface EmbeddingCreateParams {
39
39
40
40
/**
41
41
* The name of the embedding model to use.
42
+ *
43
+ * [See all of Together AI's embedding models](https://docs.together.ai/docs/serverless-models#embedding-models)
42
44
*/
43
- model : string ;
45
+ model :
46
+ | 'WhereIsAI/UAE-Large-V1'
47
+ | 'BAAI/bge-large-en-v1.5'
48
+ | 'BAAI/bge-base-en-v1.5'
49
+ | 'togethercomputer/m2-bert-80M-8k-retrieval'
50
+ | ( string & { } ) ;
44
51
}
45
52
46
53
export namespace Embeddings {
Original file line number Diff line number Diff line change @@ -34,8 +34,14 @@ export namespace ImageFile {
34
34
export interface ImageCreateParams {
35
35
/**
36
36
* The model to use for image generation.
37
+ *
38
+ * [See all of Together AI's image models](https://docs.together.ai/docs/serverless-models#image-models)
37
39
*/
38
- model : string ;
40
+ model :
41
+ | 'black-forest-labs/FLUX.1-schnell-Free'
42
+ | 'black-forest-labs/FLUX.1-schnell'
43
+ | 'black-forest-labs/FLUX.1.1-pro'
44
+ | ( string & { } ) ;
39
45
40
46
/**
41
47
* A description of the desired images. Maximum length varies by model.
Original file line number Diff line number Diff line change @@ -48,8 +48,10 @@ export interface RerankParams {
48
48
49
49
/**
50
50
* The model to be used for the rerank request.
51
+ *
52
+ * [See all of Together AI's rerank models](https://docs.together.ai/docs/serverless-models#rerank-models)
51
53
*/
52
- model : string ;
54
+ model : 'Salesforce/Llama-Rank-v1' | ( string & { } ) ;
53
55
54
56
/**
55
57
* The search query to be used for ranking.
Original file line number Diff line number Diff line change @@ -12,7 +12,7 @@ describe('resource embeddings', () => {
12
12
test ( 'create: only required params' , async ( ) => {
13
13
const responsePromise = client . embeddings . create ( {
14
14
input : 'Our solar system orbits the Milky Way galaxy at about 515,000 mph' ,
15
- model : 'togethercomputer/m2-bert-80M-8k-retrieval ' ,
15
+ model : 'WhereIsAI/UAE-Large-V1 ' ,
16
16
} ) ;
17
17
const rawResponse = await responsePromise . asResponse ( ) ;
18
18
expect ( rawResponse ) . toBeInstanceOf ( Response ) ;
@@ -26,7 +26,7 @@ describe('resource embeddings', () => {
26
26
test ( 'create: required and optional params' , async ( ) => {
27
27
const response = await client . embeddings . create ( {
28
28
input : 'Our solar system orbits the Milky Way galaxy at about 515,000 mph' ,
29
- model : 'togethercomputer/m2-bert-80M-8k-retrieval ' ,
29
+ model : 'WhereIsAI/UAE-Large-V1 ' ,
30
30
} ) ;
31
31
} ) ;
32
32
} ) ;
You can’t perform that action at this time.
0 commit comments