Class PutWatsonxRequest
- All Implemented Interfaces:
JsonpSerializable
Create an inference endpoint to perform an inference task with the
watsonxai
service. You need an IBM Cloud Databases for
Elasticsearch deployment to use the watsonxai
inference service.
You can provision one through the IBM catalog, the Cloud Databases CLI
plug-in, the Cloud Databases API, or Terraform.
When you create an inference endpoint, the associated machine learning model
is automatically deployed if it is not already running. After creating the
endpoint, wait for the model deployment to complete before using it. To
verify the deployment status, use the get trained model statistics API. Look
for "state": "fully_allocated"
in the
response and ensure that the "allocation_count"
matches the "target_allocation_count"
. Avoid creating
multiple endpoints for the same model unless required, as each endpoint
consumes significant resources.
- See Also:
-
Nested Class Summary
Nested ClassesNested classes/interfaces inherited from class co.elastic.clients.elasticsearch._types.RequestBase
RequestBase.AbstractBuilder<BuilderT extends RequestBase.AbstractBuilder<BuilderT>>
-
Field Summary
FieldsModifier and TypeFieldDescriptionstatic final JsonpDeserializer<PutWatsonxRequest>
Json deserializer forPutWatsonxRequest
static final Endpoint<PutWatsonxRequest,
PutWatsonxResponse, ErrorResponse> Endpoint "inference.put_watsonx
". -
Method Summary
Modifier and TypeMethodDescriptionstatic PutWatsonxRequest
void
serialize
(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper) Serialize this object to JSON.protected void
serializeInternal
(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper) final WatsonxServiceType
service()
Required - The type of service supported for the specified task type.final WatsonxServiceSettings
Required - Settings used to install the inference model.protected static void
final WatsonxTaskType
taskType()
Required - The task type.final String
Required - The unique identifier of the inference endpoint.Methods inherited from class co.elastic.clients.elasticsearch._types.RequestBase
toString
-
Field Details
-
_DESERIALIZER
Json deserializer forPutWatsonxRequest
-
_ENDPOINT
Endpoint "inference.put_watsonx
".
-
-
Method Details
-
of
public static PutWatsonxRequest of(Function<PutWatsonxRequest.Builder, ObjectBuilder<PutWatsonxRequest>> fn) -
service
Required - The type of service supported for the specified task type. In this case,watsonxai
.API name:
service
-
serviceSettings
Required - Settings used to install the inference model. These settings are specific to thewatsonxai
service.API name:
service_settings
-
taskType
Required - The task type. The only valid task type for the model to perform istext_embedding
.API name:
task_type
-
watsonxInferenceId
Required - The unique identifier of the inference endpoint.API name:
watsonx_inference_id
-
serialize
Serialize this object to JSON.- Specified by:
serialize
in interfaceJsonpSerializable
-
serializeInternal
-
setupPutWatsonxRequestDeserializer
protected static void setupPutWatsonxRequestDeserializer(ObjectDeserializer<PutWatsonxRequest.Builder> op)
-