Class EvaluationDatasetMetricConfig

    • Method Detail

      • taskType

        public final EvaluationTaskType taskType()

        The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.

        If the service returns an enum value that is not available in the current SDK version, taskType will return EvaluationTaskType.UNKNOWN_TO_SDK_VERSION. The raw value returned by the service is available from taskTypeAsString().

        Returns:
        The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
        See Also:
        EvaluationTaskType
      • taskTypeAsString

        public final String taskTypeAsString()

        The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.

        If the service returns an enum value that is not available in the current SDK version, taskType will return EvaluationTaskType.UNKNOWN_TO_SDK_VERSION. The raw value returned by the service is available from taskTypeAsString().

        Returns:
        The the type of task you want to evaluate for your evaluation job. This applies only to model evaluation jobs and is ignored for knowledge base evaluation jobs.
        See Also:
        EvaluationTaskType
      • dataset

        public final EvaluationDataset dataset()

        Specifies the prompt dataset.

        Returns:
        Specifies the prompt dataset.
      • hasMetricNames

        public final boolean hasMetricNames()
        For responses, this returns true if the service returned a value for the MetricNames property. This DOES NOT check that the value is non-empty (for which, you should check the isEmpty() method on the property). This is useful because the SDK will never return a null collection or map, but you may need to differentiate between the service returning nothing (or null) and the service returning an empty collection or map. For requests, this returns true if a value for the property was specified in the request builder, and false if a value was not specified.
      • metricNames

        public final List<String> metricNames()

        The names of the metrics you want to use for your evaluation job.

        For knowledge base evaluation jobs that evaluate retrieval only, valid values are " Builtin.ContextRelevance", "Builtin.ContextCoverage".

        For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are " Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness", " Builtin.LogicalCoherence", "Builtin.Faithfulness", "Builtin.Harmfulness ", "Builtin.Stereotyping", "Builtin.Refusal".

        For automated model evaluation jobs, valid values are "Builtin.Accuracy", " Builtin.Robustness", and "Builtin.Toxicity ". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", " Builtin.Completeness", "Builtin.Faithfulness", "Builtin.Helpfulness", " Builtin.Coherence", "Builtin.Relevance", "Builtin.FollowingInstructions ", "Builtin.ProfessionalStyleAndTone ", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge " Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal".

        For human-based model evaluation jobs, the list of strings must match the name parameter specified in HumanEvaluationCustomMetric.

        Attempts to modify the collection returned by this method will result in an UnsupportedOperationException.

        This method will never return null. If you would like to know whether the service returned this field (so that you can differentiate between null and empty), you can use the hasMetricNames() method.

        Returns:
        The names of the metrics you want to use for your evaluation job.

        For knowledge base evaluation jobs that evaluate retrieval only, valid values are " Builtin.ContextRelevance", "Builtin.ContextCoverage".

        For knowledge base evaluation jobs that evaluate retrieval with response generation, valid values are " Builtin.Correctness", "Builtin.Completeness", "Builtin.Helpfulness ", "Builtin.LogicalCoherence", "Builtin.Faithfulness", " Builtin.Harmfulness", "Builtin.Stereotyping", "Builtin.Refusal".

        For automated model evaluation jobs, valid values are "Builtin.Accuracy", " Builtin.Robustness", and "Builtin.Toxicity ". In model evaluation jobs that use a LLM as judge you can specify "Builtin.Correctness", " Builtin.Completeness", "Builtin.Faithfulness", " Builtin.Helpfulness", "Builtin.Coherence", "Builtin.Relevance", " Builtin.FollowingInstructions", "Builtin.ProfessionalStyleAndTone ", You can also specify the following responsible AI related metrics only for model evaluation job that use a LLM as judge " Builtin.Harmfulness", "Builtin.Stereotyping", and "Builtin.Refusal ".

        For human-based model evaluation jobs, the list of strings must match the name parameter specified in HumanEvaluationCustomMetric.

      • hashCode

        public final int hashCode()
        Overrides:
        hashCode in class Object
      • equals

        public final boolean equals​(Object obj)
        Overrides:
        equals in class Object
      • toString

        public final String toString()
        Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be redacted from this string using a placeholder value.
        Overrides:
        toString in class Object
      • getValueForField

        public final <T> Optional<T> getValueForField​(String fieldName,
                                                      Class<T> clazz)