Class EvaluationDatasetMetricConfig
- java.lang.Object
-
- software.amazon.awssdk.services.bedrock.model.EvaluationDatasetMetricConfig
-
- All Implemented Interfaces:
Serializable,SdkPojo,ToCopyableBuilder<EvaluationDatasetMetricConfig.Builder,EvaluationDatasetMetricConfig>
@Generated("software.amazon.awssdk:codegen") public final class EvaluationDatasetMetricConfig extends Object implements SdkPojo, Serializable, ToCopyableBuilder<EvaluationDatasetMetricConfig.Builder,EvaluationDatasetMetricConfig>
Defines the prompt datasets, built-in metric names and custom metric names, and the task type.
- See Also:
- Serialized Form
-
-
Nested Class Summary
Nested Classes Modifier and Type Class Description static interfaceEvaluationDatasetMetricConfig.Builder
-
Method Summary
All Methods Static Methods Instance Methods Concrete Methods Modifier and Type Method Description static EvaluationDatasetMetricConfig.Builderbuilder()EvaluationDatasetdataset()Specifies the prompt dataset.booleanequals(Object obj)booleanequalsBySdkFields(Object obj)<T> Optional<T>getValueForField(String fieldName, Class<T> clazz)inthashCode()booleanhasMetricNames()For responses, this returns true if the service returned a value for the MetricNames property.List<String>metricNames()The names of the metrics you want to use for your evaluation job.Map<String,SdkField<?>>sdkFieldNameToField()List<SdkField<?>>sdkFields()static Class<? extends EvaluationDatasetMetricConfig.Builder>serializableBuilderClass()EvaluationTaskTypetaskType()The the type of task you want to evaluate for your evaluation job.StringtaskTypeAsString()The the type of task you want to evaluate for your evaluation job.EvaluationDatasetMetricConfig.BuildertoBuilder()StringtoString()Returns a string representation of this object.-
Methods inherited from class java.lang.Object
clone, finalize, getClass, notify, notifyAll, wait, wait, wait
-
Methods inherited from interface software.amazon.awssdk.utils.builder.ToCopyableBuilder
copy
-
-
-
-
Method Detail
-
taskType
public final EvaluationTaskType taskType()
The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
If the service returns an enum value that is not available in the current SDK version,
taskTypewill returnEvaluationTaskType.UNKNOWN_TO_SDK_VERSION. The raw value returned by the service is available fromtaskTypeAsString().- Returns:
- The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
- See Also:
EvaluationTaskType
-
taskTypeAsString
public final String taskTypeAsString()
The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
If the service returns an enum value that is not available in the current SDK version,
taskTypewill returnEvaluationTaskType.UNKNOWN_TO_SDK_VERSION. The raw value returned by the service is available fromtaskTypeAsString().- Returns:
- The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
- See Also:
EvaluationTaskType
-
dataset
public final EvaluationDataset dataset()
Specifies the prompt dataset.
- Returns:
- Specifies the prompt dataset.
-
hasMetricNames
public final boolean hasMetricNames()
For responses, this returns true if the service returned a value for the MetricNames property. This DOES NOT check that the value is non-empty (for which, you should check theisEmpty()method on the property). This is useful because the SDK will never return a null collection or map, but you may need to differentiate between the service returning nothing (or null) and the service returning an empty collection or map. For requests, this returns true if a value for the property was specified in the request builder, and false if a value was not specified.
-
metricNames
public final List<String> metricNames()
The names of the metrics you want to use for your evaluation job.
For knowledge base evaluation jobs that evaluate retrieval only, valid values are "
Builtin.ContextRelevance", "Builtin.ContextCoverage".For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are "
Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", "Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".For automated model evaluation jobs, valid values are "
Builtin.Accuracy", "Builtin.Robustness", and "Builtin.Toxicity". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", "Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge "Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".For human-based model evaluation jobs, the list of strings must match the
nameparameter specified inHumanEvaluationCustomMetric.Attempts to modify the collection returned by this method will result in an UnsupportedOperationException.
This method will never return null. If you would like to know whether the service returned this field (so that you can differentiate between null and empty), you can use the
hasMetricNames()method.- Returns:
- The names of the metrics you want to use for your evaluation job.
For knowledge base evaluation jobs that evaluate retrieval only, valid values are "
Builtin.ContextRelevance", "Builtin.ContextCoverage".For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are "
Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", "Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".For automated model evaluation jobs, valid values are "
Builtin.Accuracy", "Builtin.Robustness", and "Builtin.Toxicity". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", "Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge "Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".For human-based model evaluation jobs, the list of strings must match the
nameparameter specified inHumanEvaluationCustomMetric.
-
toBuilder
public EvaluationDatasetMetricConfig.Builder toBuilder()
- Specified by:
toBuilderin interfaceToCopyableBuilder<EvaluationDatasetMetricConfig.Builder,EvaluationDatasetMetricConfig>
-
builder
public static EvaluationDatasetMetricConfig.Builder builder()
-
serializableBuilderClass
public static Class<? extends EvaluationDatasetMetricConfig.Builder> serializableBuilderClass()
-
equalsBySdkFields
public final boolean equalsBySdkFields(Object obj)
- Specified by:
equalsBySdkFieldsin interfaceSdkPojo
-
toString
public final String toString()
Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be redacted from this string using a placeholder value.
-
sdkFieldNameToField
public final Map<String,SdkField<?>> sdkFieldNameToField()
- Specified by:
sdkFieldNameToFieldin interfaceSdkPojo
-
-