Interface InferenceConfiguration.Builder
-
- All Superinterfaces:
Buildable
,CopyableBuilder<InferenceConfiguration.Builder,InferenceConfiguration>
,SdkBuilder<InferenceConfiguration.Builder,InferenceConfiguration>
,SdkPojo
- Enclosing class:
- InferenceConfiguration
public static interface InferenceConfiguration.Builder extends SdkPojo, CopyableBuilder<InferenceConfiguration.Builder,InferenceConfiguration>
-
-
Method Summary
All Methods Instance Methods Abstract Methods Modifier and Type Method Description InferenceConfiguration.Builder
maxTokens(Integer maxTokens)
The maximum number of tokens to allow in the generated response.InferenceConfiguration.Builder
stopSequences(String... stopSequences)
A list of stop sequences.InferenceConfiguration.Builder
stopSequences(Collection<String> stopSequences)
A list of stop sequences.InferenceConfiguration.Builder
temperature(Float temperature)
The likelihood of the model selecting higher-probability options while generating a response.InferenceConfiguration.Builder
topP(Float topP)
The percentage of most-likely candidates that the model considers for the next token.-
Methods inherited from interface software.amazon.awssdk.utils.builder.CopyableBuilder
copy
-
Methods inherited from interface software.amazon.awssdk.utils.builder.SdkBuilder
applyMutation, build
-
Methods inherited from interface software.amazon.awssdk.core.SdkPojo
equalsBySdkFields, sdkFieldNameToField, sdkFields
-
-
-
-
Method Detail
-
maxTokens
InferenceConfiguration.Builder maxTokens(Integer maxTokens)
The maximum number of tokens to allow in the generated response. The default value is the maximum allowed value for the model that you are using. For more information, see Inference parameters for foundation models.
- Parameters:
maxTokens
- The maximum number of tokens to allow in the generated response. The default value is the maximum allowed value for the model that you are using. For more information, see Inference parameters for foundation models.- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
temperature
InferenceConfiguration.Builder temperature(Float temperature)
The likelihood of the model selecting higher-probability options while generating a response. A lower value makes the model more likely to choose higher-probability options, while a higher value makes the model more likely to choose lower-probability options.
The default value is the default value for the model that you are using. For more information, see Inference parameters for foundation models.
- Parameters:
temperature
- The likelihood of the model selecting higher-probability options while generating a response. A lower value makes the model more likely to choose higher-probability options, while a higher value makes the model more likely to choose lower-probability options.The default value is the default value for the model that you are using. For more information, see Inference parameters for foundation models.
- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
topP
InferenceConfiguration.Builder topP(Float topP)
The percentage of most-likely candidates that the model considers for the next token. For example, if you choose a value of 0.8 for
topP
, the model selects from the top 80% of the probability distribution of tokens that could be next in the sequence.The default value is the default value for the model that you are using. For more information, see Inference parameters for foundation models.
- Parameters:
topP
- The percentage of most-likely candidates that the model considers for the next token. For example, if you choose a value of 0.8 fortopP
, the model selects from the top 80% of the probability distribution of tokens that could be next in the sequence.The default value is the default value for the model that you are using. For more information, see Inference parameters for foundation models.
- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
stopSequences
InferenceConfiguration.Builder stopSequences(Collection<String> stopSequences)
A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response.
- Parameters:
stopSequences
- A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response.- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
stopSequences
InferenceConfiguration.Builder stopSequences(String... stopSequences)
A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response.
- Parameters:
stopSequences
- A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response.- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
-