Source code for orion.benchmark.assessment.averagerank

#!/usr/bin/env python
# -*- coding: utf-8 -*-
Average Rank Assessment

from collections import defaultdict

from orion.benchmark.assessment.base import BaseAssess
from orion.plotting.base import rankings

[docs]class AverageRank(BaseAssess): """ Evaluate the average performance (objective value) between different search algorithms from the rank perspective at different time steps (trial number). The performance (objective value) used for a trial will the best result until the trial. """ def __init__(self, task_num=1): super(AverageRank, self).__init__(task_num=task_num)
[docs] def analysis(self, task, experiments): """ Generate a `plotly.graph_objects.Figure` to display average rankings between different search algorithms. task: str Name of the task experiments: list A list of (task_index, experiment), where task_index is the index of task to run for this assessment, and experiment is an instance of `orion.core.worker.experiment`. """ algorithm_groups = defaultdict(list) for _, exp in experiments: algorithm_name = list(exp.configuration["algorithms"].keys())[0] algorithm_groups[algorithm_name].append(exp) return rankings(algorithm_groups)