Class AzureAiStudioServiceSettings
java.lang.Object
co.elastic.clients.elasticsearch.inference.AzureAiStudioServiceSettings
- All Implemented Interfaces:
JsonpSerializable
@JsonpDeserializable
public class AzureAiStudioServiceSettings
extends Object
implements JsonpSerializable
- See Also:
-
Nested Class Summary
Nested Classes -
Field Summary
FieldsModifier and TypeFieldDescriptionstatic final JsonpDeserializer<AzureAiStudioServiceSettings>
Json deserializer forAzureAiStudioServiceSettings
-
Method Summary
Modifier and TypeMethodDescriptionfinal String
apiKey()
Required - A valid API key of your Azure AI Studio model deployment.final String
Required - The type of endpoint that is available for deployment through Azure AI Studio:token
orrealtime
.static AzureAiStudioServiceSettings
final String
provider()
Required - The model provider for your deployment.final RateLimitSetting
This setting helps to minimize the number of rate limit errors returned from Azure AI Studio.void
serialize
(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper) Serialize this object to JSON.protected void
serializeInternal
(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper) protected static void
setupAzureAiStudioServiceSettingsDeserializer
(ObjectDeserializer<AzureAiStudioServiceSettings.Builder> op) final String
target()
Required - The target URL of your Azure AI Studio model deployment.toString()
-
Field Details
-
_DESERIALIZER
Json deserializer forAzureAiStudioServiceSettings
-
-
Method Details
-
of
-
apiKey
Required - A valid API key of your Azure AI Studio model deployment. This key can be found on the overview page for your deployment in the management section of your Azure AI Studio account.IMPORTANT: You need to provide the API key only once, during the inference model creation. The get inference endpoint API does not retrieve your API key. After creating the inference model, you cannot change the associated API key. If you want to use a different API key, delete the inference model and recreate it with the same name and the updated API key.
API name:
api_key
-
endpointType
Required - The type of endpoint that is available for deployment through Azure AI Studio:token
orrealtime
. Thetoken
endpoint type is for "pay as you go" endpoints that are billed per token. Therealtime
endpoint type is for "real-time" endpoints that are billed per hour of usage.API name:
endpoint_type
-
target
Required - The target URL of your Azure AI Studio model deployment. This can be found on the overview page for your deployment in the management section of your Azure AI Studio account.API name:
target
-
provider
Required - The model provider for your deployment. Note that some providers may support only certain task types. Supported providers include:cohere
- available fortext_embedding
andcompletion
task typesdatabricks
- available forcompletion
task type onlymeta
- available forcompletion
task type onlymicrosoft_phi
- available forcompletion
task type onlymistral
- available forcompletion
task type onlyopenai
- available fortext_embedding
andcompletion
task types
API name:
provider
-
rateLimit
This setting helps to minimize the number of rate limit errors returned from Azure AI Studio. By default, theazureaistudio
service sets the number of requests allowed per minute to 240.API name:
rate_limit
-
serialize
Serialize this object to JSON.- Specified by:
serialize
in interfaceJsonpSerializable
-
serializeInternal
-
toString
-
setupAzureAiStudioServiceSettingsDeserializer
protected static void setupAzureAiStudioServiceSettingsDeserializer(ObjectDeserializer<AzureAiStudioServiceSettings.Builder> op)
-