Class Run.Builder

  • All Implemented Interfaces:

    
    public final class Run.Builder
    
                        

    A builder for Run.

    • Constructor Detail

    • Method Detail

      • id

         final Run.Builder id(String id)

        The identifier, which can be referenced in API endpoints.

      • maxCompletionTokens

         final Run.Builder maxCompletionTokens(Long maxCompletionTokens)

        The maximum number of completion tokens specified to have been used over the course of the run.

      • maxCompletionTokens

         final Run.Builder maxCompletionTokens(Long maxCompletionTokens)

        The maximum number of completion tokens specified to have been used over the course of the run.

      • maxPromptTokens

         final Run.Builder maxPromptTokens(Long maxPromptTokens)

        The maximum number of prompt tokens specified to have been used over the course of the run.

      • maxPromptTokens

         final Run.Builder maxPromptTokens(Long maxPromptTokens)

        The maximum number of prompt tokens specified to have been used over the course of the run.

      • metadata

         final Run.Builder metadata(Metadata metadata)

        Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format, and querying for objects via API or the dashboard.

        Keys are strings with a maximum length of 64 characters. Values are strings with a maximum length of 512 characters.

      • metadata

         final Run.Builder metadata(Optional<Metadata> metadata)

        Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format, and querying for objects via API or the dashboard.

        Keys are strings with a maximum length of 64 characters. Values are strings with a maximum length of 512 characters.

      • metadata

         final Run.Builder metadata(JsonField<Metadata> metadata)

        Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format, and querying for objects via API or the dashboard.

        Keys are strings with a maximum length of 64 characters. Values are strings with a maximum length of 512 characters.

      • responseFormat

         final Run.Builder responseFormat(AssistantResponseFormatOption responseFormat)

        Specifies the format that the model must output. Compatible with GPT-4o, GPT-4 * Turbo, and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

        Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the Structured Outputs guide.

        Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.

        Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

      • responseFormat

         final Run.Builder responseFormat(Optional<AssistantResponseFormatOption> responseFormat)

        Specifies the format that the model must output. Compatible with GPT-4o, GPT-4 * Turbo, and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

        Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the Structured Outputs guide.

        Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.

        Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

      • responseFormat

         final Run.Builder responseFormat(JsonField<AssistantResponseFormatOption> responseFormat)

        Specifies the format that the model must output. Compatible with GPT-4o, GPT-4 * Turbo, and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

        Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the Structured Outputs guide.

        Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.

        Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

      • responseFormat

         final Run.Builder responseFormat(ResponseFormatText responseFormatText)

        Specifies the format that the model must output. Compatible with GPT-4o, GPT-4 * Turbo, and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

        Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the Structured Outputs guide.

        Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.

        Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

      • responseFormat

         final Run.Builder responseFormat(ResponseFormatJsonObject responseFormatJsonObject)

        Specifies the format that the model must output. Compatible with GPT-4o, GPT-4 * Turbo, and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

        Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the Structured Outputs guide.

        Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.

        Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

      • responseFormat

         final Run.Builder responseFormat(ResponseFormatJsonSchema responseFormatJsonSchema)

        Specifies the format that the model must output. Compatible with GPT-4o, GPT-4 * Turbo, and all GPT-3.5 Turbo models since gpt-3.5-turbo-1106.

        Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema. Learn more in the Structured Outputs guide.

        Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.

        Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

      • status

         final Run.Builder status(RunStatus status)

        The status of the run, which can be either queued, in_progress, requires_action, cancelling, cancelled, failed, completed, incomplete, or expired.

      • status

         final Run.Builder status(JsonField<RunStatus> status)

        The status of the run, which can be either queued, in_progress, requires_action, cancelling, cancelled, failed, completed, incomplete, or expired.

      • toolChoice

         final Run.Builder toolChoice(AssistantToolChoiceOption toolChoice)

        Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.

      • toolChoice

         final Run.Builder toolChoice(Optional<AssistantToolChoiceOption> toolChoice)

        Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.

      • toolChoice

         final Run.Builder toolChoice(JsonField<AssistantToolChoiceOption> toolChoice)

        Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.

      • toolChoice

         final Run.Builder toolChoice(AssistantToolChoiceOption.Auto auto)

        none means the model will not call any tools and instead generates a message. auto means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user.

      • usage

         final Run.Builder usage(Run.Usage usage)

        Usage statistics related to the run. This value will be null if the run is not in a terminal state (i.e. in_progress, queued, etc.).

      • usage

         final Run.Builder usage(Optional<Run.Usage> usage)

        Usage statistics related to the run. This value will be null if the run is not in a terminal state (i.e. in_progress, queued, etc.).

      • usage

         final Run.Builder usage(JsonField<Run.Usage> usage)

        Usage statistics related to the run. This value will be null if the run is not in a terminal state (i.e. in_progress, queued, etc.).

      • topP

         final Run.Builder topP(Double topP)

        The nucleus sampling value used for this run. If not set, defaults to 1.

      • topP

         final Run.Builder topP(Double topP)

        The nucleus sampling value used for this run. If not set, defaults to 1.