Class AzureAiStudioServiceSettings
java.lang.Object
co.elastic.clients.elasticsearch.inference.AzureAiStudioServiceSettings
- All Implemented Interfaces:
JsonpSerializable
@JsonpDeserializable
public class AzureAiStudioServiceSettings
extends Object
implements JsonpSerializable
- See Also:
-
Nested Class Summary
Nested Classes -
Field Summary
FieldsModifier and TypeFieldDescriptionstatic final JsonpDeserializer<AzureAiStudioServiceSettings>Json deserializer forAzureAiStudioServiceSettings -
Method Summary
Modifier and TypeMethodDescriptionfinal StringapiKey()Required - A valid API key of your Azure AI Studio model deployment.final StringRequired - The type of endpoint that is available for deployment through Azure AI Studio:tokenorrealtime.static AzureAiStudioServiceSettingsfinal Stringprovider()Required - The model provider for your deployment.final RateLimitSettingThis setting helps to minimize the number of rate limit errors returned from Azure AI Studio.voidserialize(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper) Serialize this object to JSON.protected voidserializeInternal(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper) protected static voidsetupAzureAiStudioServiceSettingsDeserializer(ObjectDeserializer<AzureAiStudioServiceSettings.Builder> op) final Stringtarget()Required - The target URL of your Azure AI Studio model deployment.toString()
-
Field Details
-
_DESERIALIZER
Json deserializer forAzureAiStudioServiceSettings
-
-
Method Details
-
of
-
apiKey
Required - A valid API key of your Azure AI Studio model deployment. This key can be found on the overview page for your deployment in the management section of your Azure AI Studio account.IMPORTANT: You need to provide the API key only once, during the inference model creation. The get inference endpoint API does not retrieve your API key. After creating the inference model, you cannot change the associated API key. If you want to use a different API key, delete the inference model and recreate it with the same name and the updated API key.
API name:
api_key -
endpointType
Required - The type of endpoint that is available for deployment through Azure AI Studio:tokenorrealtime. Thetokenendpoint type is for "pay as you go" endpoints that are billed per token. Therealtimeendpoint type is for "real-time" endpoints that are billed per hour of usage.API name:
endpoint_type -
target
Required - The target URL of your Azure AI Studio model deployment. This can be found on the overview page for your deployment in the management section of your Azure AI Studio account.API name:
target -
provider
Required - The model provider for your deployment. Note that some providers may support only certain task types. Supported providers include:cohere- available fortext_embeddingandcompletiontask typesdatabricks- available forcompletiontask type onlymeta- available forcompletiontask type onlymicrosoft_phi- available forcompletiontask type onlymistral- available forcompletiontask type onlyopenai- available fortext_embeddingandcompletiontask types
API name:
provider -
rateLimit
This setting helps to minimize the number of rate limit errors returned from Azure AI Studio. By default, theazureaistudioservice sets the number of requests allowed per minute to 240.API name:
rate_limit -
serialize
Serialize this object to JSON.- Specified by:
serializein interfaceJsonpSerializable
-
serializeInternal
-
toString
-
setupAzureAiStudioServiceSettingsDeserializer
protected static void setupAzureAiStudioServiceSettingsDeserializer(ObjectDeserializer<AzureAiStudioServiceSettings.Builder> op)
-