Interface EvaluationDatasetMetricConfig.Builder
-
- All Superinterfaces:
Buildable,CopyableBuilder<EvaluationDatasetMetricConfig.Builder,EvaluationDatasetMetricConfig>,SdkBuilder<EvaluationDatasetMetricConfig.Builder,EvaluationDatasetMetricConfig>,SdkPojo
- Enclosing class:
- EvaluationDatasetMetricConfig
@Mutable @NotThreadSafe public static interface EvaluationDatasetMetricConfig.Builder extends SdkPojo, CopyableBuilder<EvaluationDatasetMetricConfig.Builder,EvaluationDatasetMetricConfig>
-
-
Method Summary
All Methods Instance Methods Abstract Methods Default Methods Modifier and Type Method Description default EvaluationDatasetMetricConfig.Builderdataset(Consumer<EvaluationDataset.Builder> dataset)Specifies the prompt dataset.EvaluationDatasetMetricConfig.Builderdataset(EvaluationDataset dataset)Specifies the prompt dataset.EvaluationDatasetMetricConfig.BuildermetricNames(String... metricNames)The names of the metrics you want to use for your evaluation job.EvaluationDatasetMetricConfig.BuildermetricNames(Collection<String> metricNames)The names of the metrics you want to use for your evaluation job.EvaluationDatasetMetricConfig.BuildertaskType(String taskType)The the type of task you want to evaluate for your evaluation job.EvaluationDatasetMetricConfig.BuildertaskType(EvaluationTaskType taskType)The the type of task you want to evaluate for your evaluation job.-
Methods inherited from interface software.amazon.awssdk.utils.builder.CopyableBuilder
copy
-
Methods inherited from interface software.amazon.awssdk.utils.builder.SdkBuilder
applyMutation, build
-
Methods inherited from interface software.amazon.awssdk.core.SdkPojo
equalsBySdkFields, sdkFieldNameToField, sdkFields
-
-
-
-
Method Detail
-
taskType
EvaluationDatasetMetricConfig.Builder taskType(String taskType)
The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
- Parameters:
taskType- The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.- Returns:
- Returns a reference to this object so that method calls can be chained together.
- See Also:
EvaluationTaskType,EvaluationTaskType
-
taskType
EvaluationDatasetMetricConfig.Builder taskType(EvaluationTaskType taskType)
The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
- Parameters:
taskType- The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.- Returns:
- Returns a reference to this object so that method calls can be chained together.
- See Also:
EvaluationTaskType,EvaluationTaskType
-
dataset
EvaluationDatasetMetricConfig.Builder dataset(EvaluationDataset dataset)
Specifies the prompt dataset.
- Parameters:
dataset- Specifies the prompt dataset.- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
dataset
default EvaluationDatasetMetricConfig.Builder dataset(Consumer<EvaluationDataset.Builder> dataset)
Specifies the prompt dataset.
This is a convenience method that creates an instance of theEvaluationDataset.Builderavoiding the need to create one manually viaEvaluationDataset.builder().When the
Consumercompletes,SdkBuilder.build()is called immediately and its result is passed todataset(EvaluationDataset).- Parameters:
dataset- a consumer that will call methods onEvaluationDataset.Builder- Returns:
- Returns a reference to this object so that method calls can be chained together.
- See Also:
dataset(EvaluationDataset)
-
metricNames
EvaluationDatasetMetricConfig.Builder metricNames(Collection<String> metricNames)
The names of the metrics you want to use for your evaluation job.
For knowledge base evaluation jobs that evaluate retrieval only, valid values are "
Builtin.ContextRelevance", "Builtin.ContextCoverage".For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are "
Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", "Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".For automated model evaluation jobs, valid values are "
Builtin.Accuracy", "Builtin.Robustness", and "Builtin.Toxicity". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", "Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge "Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".For human-based model evaluation jobs, the list of strings must match the
nameparameter specified inHumanEvaluationCustomMetric.- Parameters:
metricNames- The names of the metrics you want to use for your evaluation job.For knowledge base evaluation jobs that evaluate retrieval only, valid values are "
Builtin.ContextRelevance", "Builtin.ContextCoverage".For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are "
Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", "Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".For automated model evaluation jobs, valid values are "
Builtin.Accuracy", "Builtin.Robustness", and "Builtin.Toxicity". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", "Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge "Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".For human-based model evaluation jobs, the list of strings must match the
nameparameter specified inHumanEvaluationCustomMetric.- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
metricNames
EvaluationDatasetMetricConfig.Builder metricNames(String... metricNames)
The names of the metrics you want to use for your evaluation job.
For knowledge base evaluation jobs that evaluate retrieval only, valid values are "
Builtin.ContextRelevance", "Builtin.ContextCoverage".For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are "
Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", "Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".For automated model evaluation jobs, valid values are "
Builtin.Accuracy", "Builtin.Robustness", and "Builtin.Toxicity". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", "Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge "Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".For human-based model evaluation jobs, the list of strings must match the
nameparameter specified inHumanEvaluationCustomMetric.- Parameters:
metricNames- The names of the metrics you want to use for your evaluation job.For knowledge base evaluation jobs that evaluate retrieval only, valid values are "
Builtin.ContextRelevance", "Builtin.ContextCoverage".For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are "
Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", "Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".For automated model evaluation jobs, valid values are "
Builtin.Accuracy", "Builtin.Robustness", and "Builtin.Toxicity". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", "Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge "Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".For human-based model evaluation jobs, the list of strings must match the
nameparameter specified inHumanEvaluationCustomMetric.- Returns:
- Returns a reference to this object so that method calls can be chained together.
-
-