普通文本  |  307行  |  13.1 KB

# Copyright 2016 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Computes the metrics for functions, Chrome OS components and benchmarks."""

from collections import defaultdict


def ComputeDistanceForFunction(child_functions_statistics_sample,
                               child_functions_statistics_reference):
  """Computes the distance metric for a function.

  Args:
    child_functions_statistics_sample: A dict that has as a key the name of a
      function and as a value the inclusive count fraction. The keys are
      the child functions of a sample parent function.
    child_functions_statistics_reference: A dict that has as a key the name of
      a function and as a value the inclusive count fraction. The keys are
      the child functions of a reference parent function.

  Returns:
    A float value representing the sum of inclusive count fraction
    differences of pairs of common child functions. If a child function is
    present in a single data set, then we consider the missing inclusive
    count fraction as 0. This value describes the difference in behaviour
    between a sample and the reference parent function.
  """
  # We initialize the distance with a small value to avoid the further
  # division by zero.
  distance = 1.0

  for child_function, inclusive_count_fraction_reference in \
      child_functions_statistics_reference.iteritems():
    inclusive_count_fraction_sample = 0.0

    if child_function in child_functions_statistics_sample:
      inclusive_count_fraction_sample = \
          child_functions_statistics_sample[child_function]
    distance += \
        abs(inclusive_count_fraction_sample -
            inclusive_count_fraction_reference)

  for child_function, inclusive_count_fraction_sample in \
      child_functions_statistics_sample.iteritems():
    if child_function not in child_functions_statistics_reference:
      distance += inclusive_count_fraction_sample

  return distance


def ComputeScoreForFunction(distance, reference_fraction, sample_fraction):
  """Computes the score for a function.

  Args:
    distance: A float value representing the difference in behaviour between
      the sample and the reference function.
    reference_fraction: A float value representing the inclusive count
      fraction of the reference function.
    sample_fraction: A float value representing the inclusive count
      fraction of the sample function.

  Returns:
    A float value representing the score of the function.
  """
  return reference_fraction * sample_fraction / distance


def ComputeMetricsForComponents(cwp_function_groups, function_metrics):
  """Computes the metrics for a set of Chrome OS components.

  For every Chrome OS group, we compute the number of functions matching the
  group, the cumulative and average score, the cumulative and average distance
  of all those functions. A function matches a group if the path of the file
  containing its definition contains the common path describing the group.

  Args:
    cwp_function_groups: A dict having as a key the name of the group and as a
      value a common path describing the group.
    function_metrics: A dict having as a key the name of the function and the
      name of the file where it is declared concatenated by a ',', and as a
      value a tuple containing the distance and the score metrics.

  Returns:
    A dict containing as a key the name of the group and as a value a tuple
    with the group file path, the number of functions matching the group,
    the cumulative and average score, cumulative and average distance of all
    those functions.
  """
  function_groups_metrics = defaultdict(lambda: (0, 0.0, 0.0, 0.0, 0.0))

  for function_key, metric in function_metrics.iteritems():
    _, function_file = function_key.split(',')

    for group, common_path in cwp_function_groups:
      if common_path not in function_file:
        continue

      function_distance = metric[0]
      function_score = metric[1]
      group_statistic = function_groups_metrics[group]

      function_count = group_statistic[1] + 1
      function_distance_cum = function_distance + group_statistic[2]
      function_distance_avg = function_distance_cum / float(function_count)
      function_score_cum = function_score + group_statistic[4]
      function_score_avg = function_score_cum / float(function_count)

      function_groups_metrics[group] = \
          (common_path,
           function_count,
           function_distance_cum,
           function_distance_avg,
           function_score_cum,
           function_score_avg)
      break

  return function_groups_metrics


def ComputeMetricsForBenchmark(function_metrics):
  function_count = len(function_metrics.keys())
  distance_cum = 0.0
  distance_avg = 0.0
  score_cum = 0.0
  score_avg = 0.0

  for distance, score in function_metrics.values():
    distance_cum += distance
    score_cum += score

  distance_avg = distance_cum / float(function_count)
  score_avg = score_cum / float(function_count)
  return function_count, distance_cum, distance_avg, score_cum, score_avg


def ComputeFunctionCountForBenchmarkSet(set_function_metrics, cwp_functions,
                                        metric_string):
  """Computes the function count metric pair for the benchmark set.

     For the function count metric, we count the unique functions covered by the
     set of benchmarks. We compute the fraction of unique functions out
     of the amount of CWP functions given.

     We compute also the same metric pair for every group from the keys of the
     set_function_metrics dict.

  Args:
    set_function_metrics: A list of dicts having as a key the name of a group
      and as value a list of functions that match the given group.
    cwp_functions: A dict having as a key the name of the groups and as a value
      the list of CWP functions that match an individual group.
    metric_string: A tuple of strings that will be mapped to the tuple of metric
      values in the returned function group dict. This is done for convenience
      for the JSON output.

  Returns:
    A tuple with the metric pair and a dict with the group names and values
    of the metric pair. The first value of the metric pair represents the
    function count and the second value the function count fraction.
    The dict has as a key the name of the group and as a value a dict that
    maps the metric_string  to the values of the metric pair of the group.
  """
  cwp_functions_count = sum(len(functions)
                            for functions in cwp_functions.itervalues())
  set_groups_functions = defaultdict(set)
  for benchmark_function_metrics in set_function_metrics:
    for group_name in benchmark_function_metrics:
      set_groups_functions[group_name] |= \
          set(benchmark_function_metrics[group_name])

  set_groups_functions_count = {}
  set_functions_count = 0
  for group_name, functions \
      in set_groups_functions.iteritems():
    set_group_functions_count = len(functions)
    if group_name in cwp_functions:
      set_groups_functions_count[group_name] = {
          metric_string[0]: set_group_functions_count,
          metric_string[1]:
          set_group_functions_count / float(len(cwp_functions[group_name]))}
    else:
      set_groups_functions_count[group_name] = \
          {metric_string[0]: set_group_functions_count, metric_string[1]: 0.0}
    set_functions_count += set_group_functions_count

  set_functions_count_fraction = \
      set_functions_count / float(cwp_functions_count)
  return (set_functions_count, set_functions_count_fraction), \
      set_groups_functions_count


def ComputeDistanceForBenchmarkSet(set_function_metrics, cwp_functions,
                                   metric_string):
  """Computes the distance variation metric pair for the benchmark set.

     For the distance variation metric, we compute the sum of the distance
     variations of the functions covered by a set of benchmarks.
     We define the distance variation as the difference between the distance
     value of a functions and the ideal distance value (1.0).
     If a function appears in multiple common functions files, we consider
     only the minimum value. We compute also the distance variation per
     function.

     In addition, we compute also the same metric pair for every group from
     the keys of the set_function_metrics dict.

  Args:
    set_function_metrics: A list of dicts having as a key the name of a group
      and as value a list of functions that match the given group.
    cwp_functions: A dict having as a key the name of the groups and as a value
      the list of CWP functions that match an individual group.
    metric_string: A tuple of strings that will be mapped to the tuple of metric
      values in the returned function group dict. This is done for convenience
      for the JSON output.

  Returns:
    A tuple with the metric pair and a dict with the group names and values
    of the metric pair. The first value of the metric pair represents the
    distance variation per function and the second value the distance variation.
    The dict has as a key the name of the group and as a value a dict that
    maps the metric_string to the values of the metric pair of the group.
  """
  set_unique_functions = defaultdict(lambda: defaultdict(lambda: float('inf')))
  set_function_count = 0
  total_distance_variation = 0.0
  for benchmark_function_metrics in set_function_metrics:
    for group_name in benchmark_function_metrics:
      for function_key, metrics in \
          benchmark_function_metrics[group_name].iteritems():
        previous_distance = \
            set_unique_functions[group_name][function_key]
        min_distance = min(metrics[0], previous_distance)
        set_unique_functions[group_name][function_key] = min_distance
  groups_distance_variations = defaultdict(lambda: (0.0, 0.0))
  for group_name, functions_distances in set_unique_functions.iteritems():
    group_function_count = len(functions_distances)
    group_distance_variation = \
        sum(functions_distances.itervalues()) - float(group_function_count)
    total_distance_variation += group_distance_variation
    set_function_count += group_function_count
    groups_distance_variations[group_name] = \
        {metric_string[0]:
         group_distance_variation / float(group_function_count),
         metric_string[1]: group_distance_variation}

  return (total_distance_variation / set_function_count,
          total_distance_variation), groups_distance_variations


def ComputeScoreForBenchmarkSet(set_function_metrics, cwp_functions,
                                metric_string):
  """Computes the function count metric pair for the benchmark set.

     For the score metric, we compute the sum of the scores of the functions
     from a set of benchmarks. If a function appears in multiple common
     functions files, we consider only the maximum value. We compute also the
     fraction of this sum from the sum of all the scores of the functions from
     the CWP data covering the given groups, in the ideal case (the ideal
     score of a function is 1.0).

     In addition, we compute the same metric pair for every group from the
     keys of the set_function_metrics dict.

  Args:
    set_function_metrics: A list of dicts having as a key the name of a group
      and as value a list of functions that match the given group.
    cwp_functions: A dict having as a key the name of the groups and as a value
      the list of CWP functions that match an individual group.
    metric_string: A tuple of strings that will be mapped to the tuple of metric
      values in the returned function group dict. This is done for convenience
      for the JSON output.

  Returns:
    A tuple with the metric pair and a dict with the group names and values
    of the metric pair. The first value of the pair is the fraction of the sum
    of the scores from the ideal case and the second value represents the
    sum of scores of the functions. The dict has as a key the name of the group
    and as a value a dict that maps the metric_string to the values of the
    metric pair of the group.
  """
  cwp_functions_count = sum(len(functions)
                            for functions in cwp_functions.itervalues())
  set_unique_functions = defaultdict(lambda: defaultdict(lambda: 0.0))
  total_score = 0.0

  for benchmark_function_metrics in set_function_metrics:
    for group_name in benchmark_function_metrics:
      for function_key, metrics in \
          benchmark_function_metrics[group_name].iteritems():
        previous_score = \
            set_unique_functions[group_name][function_key]
        max_score = max(metrics[1], previous_score)
        set_unique_functions[group_name][function_key] = max_score

  groups_scores = defaultdict(lambda: (0.0, 0.0))

  for group_name, function_scores in set_unique_functions.iteritems():
    group_function_count = float(len(cwp_functions[group_name]))
    group_score = sum(function_scores.itervalues())
    total_score += group_score
    groups_scores[group_name] = {
        metric_string[0]: group_score / group_function_count,
        metric_string[1]: group_score
    }

  return (total_score / cwp_functions_count, total_score), groups_scores