From 4b8321663e9e603c526fb758732341b2b1d183c5 Mon Sep 17 00:00:00 2001 From: speakeasybot Date: Thu, 24 Oct 2024 13:22:13 +0000 Subject: [PATCH] ci: regenerated with OpenAPI Doc , Speakeasy CLI 1.421.1 --- .speakeasy/gen.lock | 8 ++++---- .speakeasy/gen.yaml | 2 +- .speakeasy/workflow.lock | 10 +++++----- RELEASES.md | 12 +++++++++++- .../models/components/createchatcompletionrequest.md | 1 - docs/models/components/createrunrequest.md | 1 - docs/models/components/createthreadandrunrequest.md | 1 - docs/models/components/runobject.md | 1 - jsr.json | 2 +- package-lock.json | 4 ++-- package.json | 2 +- src/lib/config.ts | 4 ++-- src/models/components/createchatcompletionrequest.ts | 9 --------- src/models/components/createrunrequest.ts | 9 --------- src/models/components/createthreadandrunrequest.ts | 9 --------- src/models/components/runobject.ts | 9 --------- 16 files changed, 27 insertions(+), 57 deletions(-) diff --git a/.speakeasy/gen.lock b/.speakeasy/gen.lock index 62dac71..6599165 100644 --- a/.speakeasy/gen.lock +++ b/.speakeasy/gen.lock @@ -1,12 +1,12 @@ lockVersion: 2.0.0 id: 61a4ac00-e7ed-4a10-9e0f-47f8aef9702c management: - docChecksum: d96ad338ba3c4b1901b5d92a729b63d9 + docChecksum: c58ce1a8f4761676d0ec62b40ce3b5af docVersion: 2.3.0 - speakeasyVersion: 1.419.2 + speakeasyVersion: 1.421.1 generationVersion: 2.438.15 - releaseVersion: 0.0.4 - configChecksum: bcf34b7256bf00bee4ddd1dba527c05d + releaseVersion: 0.0.5 + configChecksum: 20495d52a63fb99b59bfcbbc62b92c6b repoURL: https://github.com/argotdev/speakeasy-sdk.git installationURL: https://github.com/argotdev/speakeasy-sdk published: true diff --git a/.speakeasy/gen.yaml b/.speakeasy/gen.yaml index 9d8e6a2..b94cfaa 100644 --- a/.speakeasy/gen.yaml +++ b/.speakeasy/gen.yaml @@ -15,7 +15,7 @@ generation: auth: oAuth2ClientCredentialsEnabled: false typescript: - version: 0.0.4 + version: 0.0.5 additionalDependencies: dependencies: {} devDependencies: {} diff --git a/.speakeasy/workflow.lock b/.speakeasy/workflow.lock index d5db13a..1ce5a99 100644 --- a/.speakeasy/workflow.lock +++ b/.speakeasy/workflow.lock @@ -1,9 +1,9 @@ -speakeasyVersion: 1.419.2 +speakeasyVersion: 1.421.1 sources: ArgotOpenAi-OAS: sourceNamespace: argot-open-ai-oas - sourceRevisionDigest: sha256:a14f67eb2cabf63f73303b532aa095227bb699b17889ac14e3c5c00ffe2122dc - sourceBlobDigest: sha256:bbab06460b3d531ea46c3b4e1c4b378dc258d69d352cb96aa91c6ab7f0427c5f + sourceRevisionDigest: sha256:57762a25bf2b9db92394f493016221caac15a961a8b0d0daf781fbf169cf11de + sourceBlobDigest: sha256:1ac004496342b737f8cbb62c2de296ce98c3851f1082d3b6d73d7f312b6d247e tags: - latest - main @@ -11,8 +11,8 @@ targets: argot-open-ai: source: ArgotOpenAi-OAS sourceNamespace: argot-open-ai-oas - sourceRevisionDigest: sha256:a14f67eb2cabf63f73303b532aa095227bb699b17889ac14e3c5c00ffe2122dc - sourceBlobDigest: sha256:bbab06460b3d531ea46c3b4e1c4b378dc258d69d352cb96aa91c6ab7f0427c5f + sourceRevisionDigest: sha256:57762a25bf2b9db92394f493016221caac15a961a8b0d0daf781fbf169cf11de + sourceBlobDigest: sha256:1ac004496342b737f8cbb62c2de296ce98c3851f1082d3b6d73d7f312b6d247e workflow: workflowVersion: 1.0.0 speakeasyVersion: latest diff --git a/RELEASES.md b/RELEASES.md index 3da2da3..acaaf65 100644 --- a/RELEASES.md +++ b/RELEASES.md @@ -26,4 +26,14 @@ Based on: ### Generated - [typescript v0.0.4] . ### Releases -- [NPM v0.0.4] https://www.npmjs.com/package/argot-open-ai/v/0.0.4 - . \ No newline at end of file +- [NPM v0.0.4] https://www.npmjs.com/package/argot-open-ai/v/0.0.4 - . + +## 2024-10-24 13:20:55 +### Changes +Based on: +- OpenAPI Doc +- Speakeasy CLI 1.421.1 (2.438.15) https://github.com/speakeasy-api/speakeasy +### Generated +- [typescript v0.0.5] . +### Releases +- [NPM v0.0.5] https://www.npmjs.com/package/argot-open-ai/v/0.0.5 - . \ No newline at end of file diff --git a/docs/models/components/createchatcompletionrequest.md b/docs/models/components/createchatcompletionrequest.md index 18b4e36..81872a4 100644 --- a/docs/models/components/createchatcompletionrequest.md +++ b/docs/models/components/createchatcompletionrequest.md @@ -49,7 +49,6 @@ let value: CreateChatCompletionRequest = { | `topP` | *number* | :heavy_minus_sign: | An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered.

We generally recommend altering this or `temperature` but not both.
| 1 | | `tools` | [components.ChatCompletionTool](../../models/components/chatcompletiontool.md)[] | :heavy_minus_sign: | A list of tools the model may call. Currently, only functions are supported as a tool. Use this to provide a list of functions the model may generate JSON inputs for. A max of 128 functions are supported.
| | | `toolChoice` | *components.ChatCompletionToolChoiceOption* | :heavy_minus_sign: | Controls which (if any) tool is called by the model.
`none` means the model will not call any tool and instead generates a message.
`auto` means the model can pick between generating a message or calling one or more tools.
`required` means the model must call one or more tools.
Specifying a particular tool via `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool.

`none` is the default when no tools are present. `auto` is the default if tools are present.
| | -| `parallelToolCalls` | *boolean* | :heavy_minus_sign: | Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. | | | `user` | *string* | :heavy_minus_sign: | A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. [Learn more](/docs/guides/safety-best-practices/end-user-ids).
| user-1234 | | ~~`functionCall`~~ | *components.CreateChatCompletionRequestFunctionCall* | :heavy_minus_sign: | : warning: ** DEPRECATED **: This will be removed in a future release, please migrate away from it as soon as possible.

Deprecated in favor of `tool_choice`.

Controls which (if any) function is called by the model.
`none` means the model will not call a function and instead generates a message.
`auto` means the model can pick between generating a message or calling a function.
Specifying a particular function via `{"name": "my_function"}` forces the model to call that function.

`none` is the default when no functions are present. `auto` is the default if functions are present.
| | | ~~`functions`~~ | [components.ChatCompletionFunctions](../../models/components/chatcompletionfunctions.md)[] | :heavy_minus_sign: | : warning: ** DEPRECATED **: This will be removed in a future release, please migrate away from it as soon as possible.

Deprecated in favor of `tools`.

A list of functions the model may generate JSON inputs for.
| | \ No newline at end of file diff --git a/docs/models/components/createrunrequest.md b/docs/models/components/createrunrequest.md index 938fac8..0ff9feb 100644 --- a/docs/models/components/createrunrequest.md +++ b/docs/models/components/createrunrequest.md @@ -31,5 +31,4 @@ let value: CreateRunRequest = { | `maxCompletionTokens` | *number* | :heavy_minus_sign: | The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status `incomplete`. See `incomplete_details` for more info.
| | | `truncationStrategy` | [components.TruncationObject](../../models/components/truncationobject.md) | :heavy_minus_sign: | Controls for how a thread will be truncated prior to the run. Use this to control the intial context window of the run. | | | `toolChoice` | *components.AssistantsApiToolChoiceOption* | :heavy_minus_sign: | Controls which (if any) tool is called by the model.
`none` means the model will not call any tools and instead generates a message.
`auto` is the default value and means the model can pick between generating a message or calling one or more tools.
`required` means the model must call one or more tools before responding to the user.
Specifying a particular tool like `{"type": "file_search"}` or `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool.
| | -| `parallelToolCalls` | *boolean* | :heavy_minus_sign: | Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. | | | `responseFormat` | *components.AssistantsApiResponseFormatOption* | :heavy_minus_sign: | Specifies the format that the model must output. Compatible with [GPT-4o](/docs/models/gpt-4o), [GPT-4 Turbo](/docs/models/gpt-4-turbo-and-gpt-4), and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.

Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the [Structured Outputs guide](/docs/guides/structured-outputs).

Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the message the model generates is valid JSON.

**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.
| | \ No newline at end of file diff --git a/docs/models/components/createthreadandrunrequest.md b/docs/models/components/createthreadandrunrequest.md index 6c4b5c1..4e49a35 100644 --- a/docs/models/components/createthreadandrunrequest.md +++ b/docs/models/components/createthreadandrunrequest.md @@ -31,5 +31,4 @@ let value: CreateThreadAndRunRequest = { | `maxCompletionTokens` | *number* | :heavy_minus_sign: | The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status `incomplete`. See `incomplete_details` for more info.
| | | `truncationStrategy` | [components.TruncationObject](../../models/components/truncationobject.md) | :heavy_minus_sign: | Controls for how a thread will be truncated prior to the run. Use this to control the intial context window of the run. | | | `toolChoice` | *components.AssistantsApiToolChoiceOption* | :heavy_minus_sign: | Controls which (if any) tool is called by the model.
`none` means the model will not call any tools and instead generates a message.
`auto` is the default value and means the model can pick between generating a message or calling one or more tools.
`required` means the model must call one or more tools before responding to the user.
Specifying a particular tool like `{"type": "file_search"}` or `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool.
| | -| `parallelToolCalls` | *boolean* | :heavy_minus_sign: | Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. | | | `responseFormat` | *components.AssistantsApiResponseFormatOption* | :heavy_minus_sign: | Specifies the format that the model must output. Compatible with [GPT-4o](/docs/models/gpt-4o), [GPT-4 Turbo](/docs/models/gpt-4-turbo-and-gpt-4), and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.

Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the [Structured Outputs guide](/docs/guides/structured-outputs).

Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the message the model generates is valid JSON.

**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.
| | \ No newline at end of file diff --git a/docs/models/components/runobject.md b/docs/models/components/runobject.md index 80110cd..cf46de4 100644 --- a/docs/models/components/runobject.md +++ b/docs/models/components/runobject.md @@ -98,5 +98,4 @@ let value: RunObject = { | `maxCompletionTokens` | *number* | :heavy_check_mark: | The maximum number of completion tokens specified to have been used over the course of the run.
| | `truncationStrategy` | [components.TruncationObject](../../models/components/truncationobject.md) | :heavy_check_mark: | Controls for how a thread will be truncated prior to the run. Use this to control the intial context window of the run. | | `toolChoice` | *components.AssistantsApiToolChoiceOption* | :heavy_check_mark: | Controls which (if any) tool is called by the model.
`none` means the model will not call any tools and instead generates a message.
`auto` is the default value and means the model can pick between generating a message or calling one or more tools.
`required` means the model must call one or more tools before responding to the user.
Specifying a particular tool like `{"type": "file_search"}` or `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool.
| -| `parallelToolCalls` | *boolean* | :heavy_minus_sign: | Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. | | `responseFormat` | *components.AssistantsApiResponseFormatOption* | :heavy_check_mark: | Specifies the format that the model must output. Compatible with [GPT-4o](/docs/models/gpt-4o), [GPT-4 Turbo](/docs/models/gpt-4-turbo-and-gpt-4), and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.

Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the [Structured Outputs guide](/docs/guides/structured-outputs).

Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the message the model generates is valid JSON.

**Important:** when using JSON mode, you **must** also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if `finish_reason="length"`, which indicates the generation exceeded `max_tokens` or the conversation exceeded the max context length.
| \ No newline at end of file diff --git a/jsr.json b/jsr.json index 407c381..b3c37ed 100644 --- a/jsr.json +++ b/jsr.json @@ -2,7 +2,7 @@ { "name": "argot-open-ai", - "version": "0.0.4", + "version": "0.0.5", "exports": { ".": "./src/index.ts", "./models/errors": "./src/models/errors/index.ts", diff --git a/package-lock.json b/package-lock.json index 3bf5906..9bb9662 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1,12 +1,12 @@ { "name": "argot-open-ai", - "version": "0.0.4", + "version": "0.0.5", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "argot-open-ai", - "version": "0.0.4", + "version": "0.0.5", "devDependencies": { "@typescript-eslint/eslint-plugin": "^7.7.1", "@typescript-eslint/parser": "^7.7.1", diff --git a/package.json b/package.json index 6518e92..2033ab7 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "argot-open-ai", - "version": "0.0.4", + "version": "0.0.5", "author": "Speakeasy", "main": "./index.js", "sideEffects": false, diff --git a/src/lib/config.ts b/src/lib/config.ts index 1a4b344..228d486 100644 --- a/src/lib/config.ts +++ b/src/lib/config.ts @@ -54,7 +54,7 @@ export function serverURLFromOptions(options: SDKOptions): URL | null { export const SDK_METADATA = { language: "typescript", openapiDocVersion: "2.3.0", - sdkVersion: "0.0.4", + sdkVersion: "0.0.5", genVersion: "2.438.15", - userAgent: "speakeasy-sdk/typescript 0.0.4 2.438.15 2.3.0 argot-open-ai", + userAgent: "speakeasy-sdk/typescript 0.0.5 2.438.15 2.3.0 argot-open-ai", } as const; diff --git a/src/models/components/createchatcompletionrequest.ts b/src/models/components/createchatcompletionrequest.ts index e49cbd2..b810cae 100644 --- a/src/models/components/createchatcompletionrequest.ts +++ b/src/models/components/createchatcompletionrequest.ts @@ -332,10 +332,6 @@ export type CreateChatCompletionRequest = { * `none` is the default when no tools are present. `auto` is the default if tools are present. */ toolChoice?: ChatCompletionToolChoiceOption | undefined; - /** - * Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. - */ - parallelToolCalls?: boolean | undefined; /** * A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. [Learn more](/docs/guides/safety-best-practices/end-user-ids). * @@ -590,7 +586,6 @@ export const CreateChatCompletionRequest$inboundSchema: z.ZodType< top_p: z.nullable(z.number().default(1)), tools: z.array(ChatCompletionTool$inboundSchema).optional(), tool_choice: ChatCompletionToolChoiceOption$inboundSchema.optional(), - parallel_tool_calls: z.boolean().default(true), user: z.string().optional(), function_call: z.union([ ChatCompletionFunctionCallOption$inboundSchema, @@ -610,7 +605,6 @@ export const CreateChatCompletionRequest$inboundSchema: z.ZodType< "stream_options": "streamOptions", "top_p": "topP", "tool_choice": "toolChoice", - "parallel_tool_calls": "parallelToolCalls", "function_call": "functionCall", }); }); @@ -641,7 +635,6 @@ export type CreateChatCompletionRequest$Outbound = { top_p: number | null; tools?: Array | undefined; tool_choice?: ChatCompletionToolChoiceOption$Outbound | undefined; - parallel_tool_calls: boolean; user?: string | undefined; function_call?: | ChatCompletionFunctionCallOption$Outbound @@ -681,7 +674,6 @@ export const CreateChatCompletionRequest$outboundSchema: z.ZodType< topP: z.nullable(z.number().default(1)), tools: z.array(ChatCompletionTool$outboundSchema).optional(), toolChoice: ChatCompletionToolChoiceOption$outboundSchema.optional(), - parallelToolCalls: z.boolean().default(true), user: z.string().optional(), functionCall: z.union([ ChatCompletionFunctionCallOption$outboundSchema, @@ -701,7 +693,6 @@ export const CreateChatCompletionRequest$outboundSchema: z.ZodType< streamOptions: "stream_options", topP: "top_p", toolChoice: "tool_choice", - parallelToolCalls: "parallel_tool_calls", functionCall: "function_call", }); }); diff --git a/src/models/components/createrunrequest.ts b/src/models/components/createrunrequest.ts index ef20e64..33731b2 100644 --- a/src/models/components/createrunrequest.ts +++ b/src/models/components/createrunrequest.ts @@ -174,10 +174,6 @@ export type CreateRunRequest = { * Specifying a particular tool like `{"type": "file_search"}` or `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool. */ toolChoice?: AssistantsApiToolChoiceOption | undefined; - /** - * Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. - */ - parallelToolCalls?: boolean | undefined; /** * Specifies the format that the model must output. Compatible with [GPT-4o](/docs/models/gpt-4o), [GPT-4 Turbo](/docs/models/gpt-4-turbo-and-gpt-4), and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`. * @@ -345,7 +341,6 @@ export const CreateRunRequest$inboundSchema: z.ZodType< max_completion_tokens: z.nullable(z.number().int()).optional(), truncation_strategy: TruncationObject$inboundSchema.optional(), tool_choice: AssistantsApiToolChoiceOption$inboundSchema.optional(), - parallel_tool_calls: z.boolean().default(true), response_format: AssistantsApiResponseFormatOption$inboundSchema.optional(), }).transform((v) => { return remap$(v, { @@ -357,7 +352,6 @@ export const CreateRunRequest$inboundSchema: z.ZodType< "max_completion_tokens": "maxCompletionTokens", "truncation_strategy": "truncationStrategy", "tool_choice": "toolChoice", - "parallel_tool_calls": "parallelToolCalls", "response_format": "responseFormat", }); }); @@ -385,7 +379,6 @@ export type CreateRunRequest$Outbound = { max_completion_tokens?: number | null | undefined; truncation_strategy?: TruncationObject$Outbound | undefined; tool_choice?: AssistantsApiToolChoiceOption$Outbound | undefined; - parallel_tool_calls: boolean; response_format?: AssistantsApiResponseFormatOption$Outbound | undefined; }; @@ -421,7 +414,6 @@ export const CreateRunRequest$outboundSchema: z.ZodType< maxCompletionTokens: z.nullable(z.number().int()).optional(), truncationStrategy: TruncationObject$outboundSchema.optional(), toolChoice: AssistantsApiToolChoiceOption$outboundSchema.optional(), - parallelToolCalls: z.boolean().default(true), responseFormat: AssistantsApiResponseFormatOption$outboundSchema.optional(), }).transform((v) => { return remap$(v, { @@ -433,7 +425,6 @@ export const CreateRunRequest$outboundSchema: z.ZodType< maxCompletionTokens: "max_completion_tokens", truncationStrategy: "truncation_strategy", toolChoice: "tool_choice", - parallelToolCalls: "parallel_tool_calls", responseFormat: "response_format", }); }); diff --git a/src/models/components/createthreadandrunrequest.ts b/src/models/components/createthreadandrunrequest.ts index 44c2d67..3cd8980 100644 --- a/src/models/components/createthreadandrunrequest.ts +++ b/src/models/components/createthreadandrunrequest.ts @@ -205,10 +205,6 @@ export type CreateThreadAndRunRequest = { * Specifying a particular tool like `{"type": "file_search"}` or `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool. */ toolChoice?: AssistantsApiToolChoiceOption | undefined; - /** - * Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. - */ - parallelToolCalls?: boolean | undefined; /** * Specifies the format that the model must output. Compatible with [GPT-4o](/docs/models/gpt-4o), [GPT-4 Turbo](/docs/models/gpt-4-turbo-and-gpt-4), and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`. * @@ -531,7 +527,6 @@ export const CreateThreadAndRunRequest$inboundSchema: z.ZodType< max_completion_tokens: z.nullable(z.number().int()).optional(), truncation_strategy: TruncationObject$inboundSchema.optional(), tool_choice: AssistantsApiToolChoiceOption$inboundSchema.optional(), - parallel_tool_calls: z.boolean().default(true), response_format: AssistantsApiResponseFormatOption$inboundSchema.optional(), }).transform((v) => { return remap$(v, { @@ -542,7 +537,6 @@ export const CreateThreadAndRunRequest$inboundSchema: z.ZodType< "max_completion_tokens": "maxCompletionTokens", "truncation_strategy": "truncationStrategy", "tool_choice": "toolChoice", - "parallel_tool_calls": "parallelToolCalls", "response_format": "responseFormat", }); }); @@ -573,7 +567,6 @@ export type CreateThreadAndRunRequest$Outbound = { max_completion_tokens?: number | null | undefined; truncation_strategy?: TruncationObject$Outbound | undefined; tool_choice?: AssistantsApiToolChoiceOption$Outbound | undefined; - parallel_tool_calls: boolean; response_format?: AssistantsApiResponseFormatOption$Outbound | undefined; }; @@ -611,7 +604,6 @@ export const CreateThreadAndRunRequest$outboundSchema: z.ZodType< maxCompletionTokens: z.nullable(z.number().int()).optional(), truncationStrategy: TruncationObject$outboundSchema.optional(), toolChoice: AssistantsApiToolChoiceOption$outboundSchema.optional(), - parallelToolCalls: z.boolean().default(true), responseFormat: AssistantsApiResponseFormatOption$outboundSchema.optional(), }).transform((v) => { return remap$(v, { @@ -622,7 +614,6 @@ export const CreateThreadAndRunRequest$outboundSchema: z.ZodType< maxCompletionTokens: "max_completion_tokens", truncationStrategy: "truncation_strategy", toolChoice: "tool_choice", - parallelToolCalls: "parallel_tool_calls", responseFormat: "response_format", }); }); diff --git a/src/models/components/runobject.ts b/src/models/components/runobject.ts index 05a1274..d14d606 100644 --- a/src/models/components/runobject.ts +++ b/src/models/components/runobject.ts @@ -298,10 +298,6 @@ export type RunObject = { * Specifying a particular tool like `{"type": "file_search"}` or `{"type": "function", "function": {"name": "my_function"}}` forces the model to call that tool. */ toolChoice: AssistantsApiToolChoiceOption; - /** - * Whether to enable [parallel function calling](/docs/guides/function-calling/parallel-function-calling) during tool use. - */ - parallelToolCalls?: boolean | undefined; /** * Specifies the format that the model must output. Compatible with [GPT-4o](/docs/models/gpt-4o), [GPT-4 Turbo](/docs/models/gpt-4-turbo-and-gpt-4), and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`. * @@ -696,7 +692,6 @@ export const RunObject$inboundSchema: z.ZodType< max_completion_tokens: z.nullable(z.number().int()), truncation_strategy: TruncationObject$inboundSchema, tool_choice: AssistantsApiToolChoiceOption$inboundSchema, - parallel_tool_calls: z.boolean().default(true), response_format: AssistantsApiResponseFormatOption$inboundSchema, }).transform((v) => { return remap$(v, { @@ -716,7 +711,6 @@ export const RunObject$inboundSchema: z.ZodType< "max_completion_tokens": "maxCompletionTokens", "truncation_strategy": "truncationStrategy", "tool_choice": "toolChoice", - "parallel_tool_calls": "parallelToolCalls", "response_format": "responseFormat", }); }); @@ -752,7 +746,6 @@ export type RunObject$Outbound = { max_completion_tokens: number | null; truncation_strategy: TruncationObject$Outbound; tool_choice: AssistantsApiToolChoiceOption$Outbound; - parallel_tool_calls: boolean; response_format: AssistantsApiResponseFormatOption$Outbound; }; @@ -795,7 +788,6 @@ export const RunObject$outboundSchema: z.ZodType< maxCompletionTokens: z.nullable(z.number().int()), truncationStrategy: TruncationObject$outboundSchema, toolChoice: AssistantsApiToolChoiceOption$outboundSchema, - parallelToolCalls: z.boolean().default(true), responseFormat: AssistantsApiResponseFormatOption$outboundSchema, }).transform((v) => { return remap$(v, { @@ -815,7 +807,6 @@ export const RunObject$outboundSchema: z.ZodType< maxCompletionTokens: "max_completion_tokens", truncationStrategy: "truncation_strategy", toolChoice: "tool_choice", - parallelToolCalls: "parallel_tool_calls", responseFormat: "response_format", }); });