Class ElasticsearchServiceSettings

java.lang.Object
co.elastic.clients.elasticsearch.inference.ElasticsearchServiceSettings
All Implemented Interfaces:
JsonpSerializable

@JsonpDeserializable public class ElasticsearchServiceSettings extends Object implements JsonpSerializable
See Also:
  • Field Details

  • Method Details

    • of

    • adaptiveAllocations

      @Nullable public final AdaptiveAllocations adaptiveAllocations()
      Adaptive allocations configuration details. If enabled is true, the number of allocations of the model is set based on the current load the process gets. When the load is high, a new model allocation is automatically created, respecting the value of max_number_of_allocations if it's set. When the load is low, a model allocation is automatically removed, respecting the value of min_number_of_allocations if it's set. If enabled is true, do not set the number of allocations manually.

      API name: adaptive_allocations

    • deploymentId

      @Nullable public final String deploymentId()
      The deployment identifier for a trained model deployment. When deployment_id is used the model_id is optional.

      API name: deployment_id

    • modelId

      public final String modelId()
      Required - The name of the model to use for the inference task. It can be the ID of a built-in model (for example, .multilingual-e5-small for E5) or a text embedding model that was uploaded by using the Eland client.

      API name: model_id

    • numAllocations

      @Nullable public final Integer numAllocations()
      The total number of allocations that are assigned to the model across machine learning nodes. Increasing this value generally increases the throughput. If adaptive allocations are enabled, do not set this value because it's automatically set.

      API name: num_allocations

    • numThreads

      public final int numThreads()
      Required - The number of threads used by each model allocation during inference. This setting generally increases the speed per inference request. The inference process is a compute-bound process; threads_per_allocations must not exceed the number of available allocated processors per node. The value must be a power of 2. The maximum value is 32.

      API name: num_threads

    • serialize

      public void serialize(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper)
      Serialize this object to JSON.
      Specified by:
      serialize in interface JsonpSerializable
    • serializeInternal

      protected void serializeInternal(jakarta.json.stream.JsonGenerator generator, JsonpMapper mapper)
    • toString

      public String toString()
      Overrides:
      toString in class Object
    • setupElasticsearchServiceSettingsDeserializer

      protected static void setupElasticsearchServiceSettingsDeserializer(ObjectDeserializer<ElasticsearchServiceSettings.Builder> op)