Class ChatCompletionCreateParams.ServiceTier

  • All Implemented Interfaces:
    com.openai.core.Enum

    
    public final class ChatCompletionCreateParams.ServiceTier
     implements Enum
                        

    Specifies the latency tier to use for processing the request. This parameter is relevant for customers subscribed to the scale tier service:

    • If set to 'auto', and the Project is Scale tier enabled, the system will utilize scale tier credits until they are exhausted.

    • If set to 'auto', and the Project is not Scale tier enabled, the request will be processed using the default service tier with a lower uptime SLA and no latency guarantee.

    • If set to 'default', the request will be processed using the default service tier with a lower uptime SLA and no latency guarantee.

    • When not set, the default behavior is 'auto'.