public final class EvaluationAggregator extends Object
Modifier and Type | Method and Description |
---|---|
static <T extends Output<T>,C extends MetricContext<T>> |
argmax(EvaluationMetric<T,C> metric,
List<? extends Model<T>> models,
Dataset<T> dataset)
Calculates the argmax of a metric across the supplied models (i.e., the index of the model which performed the best).
|
static <T extends Output<T>,C extends MetricContext<T>> |
argmax(EvaluationMetric<T,C> metric,
Model<T> model,
List<? extends Dataset<T>> datasets)
Calculates the argmax of a metric across the supplied datasets.
|
static <T extends Output<T>,R extends Evaluation<T>> |
argmax(List<R> evaluations,
Function<R,Double> getter)
Calculates the argmax of a metric across the supplied evaluations.
|
static <T extends Output<T>,C extends MetricContext<T>> |
summarize(EvaluationMetric<T,C> metric,
List<? extends Model<T>> models,
Dataset<T> dataset)
Summarize performance w.r.t.
|
static <T extends Output<T>,C extends MetricContext<T>> |
summarize(EvaluationMetric<T,C> metric,
Model<T> model,
List<? extends Dataset<T>> datasets)
Summarize a model's performance w.r.t.
|
static <T extends Output<T>,R extends Evaluation<T>> |
summarize(Evaluator<T,R> evaluator,
List<? extends Model<T>> models,
Dataset<T> dataset)
Summarize performance using the supplied evaluator across several models on one dataset.
|
static <T extends Output<T>,R extends Evaluation<T>> |
summarize(Evaluator<T,R> evaluator,
Model<T> model,
List<? extends Dataset<T>> datasets)
Summarize performance according to evaluator for a single model across several datasets.
|
static <T extends Output<T>,C extends MetricContext<T>> |
summarize(List<? extends EvaluationMetric<T,C>> metrics,
Model<T> model,
Dataset<T> dataset)
Summarize model performance on dataset across several metrics.
|
static <T extends Output<T>,C extends MetricContext<T>> |
summarize(List<? extends EvaluationMetric<T,C>> metrics,
Model<T> model,
List<Prediction<T>> predictions)
Summarize model performance on dataset across several metrics.
|
static <T extends Output<T>,R extends Evaluation<T>> |
summarize(List<R> evaluations)
Summarize all fields of a list of evaluations.
|
static <T extends Output<T>,R extends Evaluation<T>> |
summarize(List<R> evaluations,
ToDoubleFunction<R> fieldGetter)
Summarize a single field of an evaluation across several evaluations.
|
public static <T extends Output<T>,C extends MetricContext<T>> DescriptiveStats summarize(EvaluationMetric<T,C> metric, List<? extends Model<T>> models, Dataset<T> dataset)
T
- The output type.C
- The context type used for this metric.metric
- The metric to summarise.models
- The models to evaluate.dataset
- The dataset to evaluate.public static <T extends Output<T>,R extends Evaluation<T>> Map<MetricID<T>,DescriptiveStats> summarize(Evaluator<T,R> evaluator, List<? extends Model<T>> models, Dataset<T> dataset)
T
- The output type.R
- The evaluation type.evaluator
- The evaluator to use.models
- The models to evaluate.dataset
- The dataset to evaluate.public static <T extends Output<T>,C extends MetricContext<T>> DescriptiveStats summarize(EvaluationMetric<T,C> metric, Model<T> model, List<? extends Dataset<T>> datasets)
T
- The output type.C
- The metric context type.metric
- The metric to evaluate.model
- The model to evaluate.datasets
- The datasets to evaluate.public static <T extends Output<T>,C extends MetricContext<T>> DescriptiveStats summarize(List<? extends EvaluationMetric<T,C>> metrics, Model<T> model, Dataset<T> dataset)
T
- The output type.C
- The metric context type.metrics
- The metrics to evaluate.model
- The model to evaluate them on.dataset
- The dataset to evaluate them on.public static <T extends Output<T>,C extends MetricContext<T>> DescriptiveStats summarize(List<? extends EvaluationMetric<T,C>> metrics, Model<T> model, List<Prediction<T>> predictions)
T
- The output type.C
- The metric context type.metrics
- The metrics to evaluate.model
- The model to evaluate them on.predictions
- The predictions to evaluate.public static <T extends Output<T>,R extends Evaluation<T>> Map<MetricID<T>,DescriptiveStats> summarize(Evaluator<T,R> evaluator, Model<T> model, List<? extends Dataset<T>> datasets)
T
- The output type.R
- The evaluation type.evaluator
- The evaluator to use.model
- The model to evaluate.datasets
- The datasets to evaluate across.public static <T extends Output<T>,R extends Evaluation<T>> Map<MetricID<T>,DescriptiveStats> summarize(List<R> evaluations)
T
- The output type.R
- The evaluation type.evaluations
- The evaluations to summarize.public static <T extends Output<T>,R extends Evaluation<T>> DescriptiveStats summarize(List<R> evaluations, ToDoubleFunction<R> fieldGetter)
T
- the type of the outputR
- the type of the evaluationevaluations
- the evaluationsfieldGetter
- the getter for the field to summarizepublic static <T extends Output<T>,C extends MetricContext<T>> com.oracle.labs.mlrg.olcut.util.Pair<Integer,Double> argmax(EvaluationMetric<T,C> metric, List<? extends Model<T>> models, Dataset<T> dataset)
T
- The output type.C
- The metric context.metric
- The metric to evaluate.models
- The models to evaluate across.dataset
- The dataset to evaluate on.public static <T extends Output<T>,C extends MetricContext<T>> com.oracle.labs.mlrg.olcut.util.Pair<Integer,Double> argmax(EvaluationMetric<T,C> metric, Model<T> model, List<? extends Dataset<T>> datasets)
T
- The output type.C
- The metric context.metric
- The metric to evaluate.model
- The model to evaluate on.datasets
- The datasets to evaluate across.public static <T extends Output<T>,R extends Evaluation<T>> com.oracle.labs.mlrg.olcut.util.Pair<Integer,Double> argmax(List<R> evaluations, Function<R,Double> getter)
T
- The output type.R
- The evaluation type.evaluations
- The evaluations.getter
- The function to extract a value from the evaluation.Copyright © 2015–2021 Oracle and/or its affiliates. All rights reserved.