Class ChatCompletionChunk

  • All Implemented Interfaces:

    
    public final class ChatCompletionChunk
    
                        

    Represents a streamed chunk of a chat completion response returned by the model, based on the provided input. Learn more.

    • Constructor Detail

    • Method Detail

      • id

         final String id()

        A unique identifier for the chat completion. Each chunk has the same ID.

      • choices

         final List<ChatCompletionChunk.Choice> choices()

        A list of chat completion choices. Can contain more than one elements if n is greater than 1. Can also be empty for the last chunk if you set stream_options: {"include_usage": true}.

      • created

         final Long created()

        The Unix timestamp (in seconds) of when the chat completion was created. Each chunk has the same timestamp.

      • model

         final String model()

        The model to generate the completion.

      • _object_

         final JsonValue _object_()

        The object type, which is always chat.completion.chunk.

        Expected to always return the following:

        JsonValue.from("chat.completion.chunk")

        However, this method can be useful for debugging and logging (e.g. if the server responded with an unexpected value).

      • serviceTier

         final Optional<ChatCompletionChunk.ServiceTier> serviceTier()

        Specifies the latency tier to use for processing the request. This parameter is relevant for customers subscribed to the scale tier service:

        • If set to 'auto', and the Project is Scale tier enabled, the system will utilize scale tier credits until they are exhausted.

        • If set to 'auto', and the Project is not Scale tier enabled, the request will be processed using the default service tier with a lower uptime SLA and no latency guarentee.

        • If set to 'default', the request will be processed using the default service tier with a lower uptime SLA and no latency guarentee.

        • If set to 'flex', the request will be processed with the Flex Processing service tier. Learn more.

        • When not set, the default behavior is 'auto'.

        When this parameter is set, the response body will include the service_tier utilized.

      • systemFingerprint

         final Optional<String> systemFingerprint()

        This fingerprint represents the backend configuration that the model runs with. Can be used in conjunction with the seed request parameter to understand when backend changes have been made that might impact determinism.

      • usage

         final Optional<CompletionUsage> usage()

        An optional field that will only be present when you set stream_options: {"include_usage": true} in your request. When present, it contains a null value except for the last chunk which contains the token usage statistics for the entire request.

        NOTE: If the stream is interrupted or cancelled, you may not receive the final usage chunk which contains the total token usage for the request.

      • _id

         final JsonField<String> _id()

        Returns the raw JSON value of id.

        Unlike id, this method doesn't throw if the JSON field has an unexpected type.

      • _model

         final JsonField<String> _model()

        Returns the raw JSON value of model.

        Unlike model, this method doesn't throw if the JSON field has an unexpected type.