Source code for ignite.contrib.metrics.average_precision

from ignite.metrics import EpochMetric

def average_precision_compute_fn(y_preds, y_targets):
        from sklearn.metrics import average_precision_score
    except ImportError:
        raise RuntimeError("This contrib module requires sklearn to be installed.")

    y_true = y_targets.numpy()
    y_pred = y_preds.numpy()
    return average_precision_score(y_true, y_pred)

[docs]class AveragePrecision(EpochMetric): """Computes Average Precision accumulating predictions and the ground-truth during an epoch and applying `sklearn.metrics.average_precision_score < sklearn.metrics.average_precision_score.html#sklearn.metrics.average_precision_score>`_ . Args: output_transform (callable, optional): a callable that is used to transform the :class:`~ignite.engine.Engine`'s `process_function`'s output into the form expected by the metric. This can be useful if, for example, you have a multi-output model and you want to compute the metric with respect to one of the outputs. AveragePrecision expects y to be comprised of 0's and 1's. y_pred must either be probability estimates or confidence values. To apply an activation to y_pred, use output_transform as shown below: .. code-block:: python def activated_output_transform(output): y_pred, y = output y_pred = torch.softmax(y_pred, dim=1) return y_pred, y avg_precision = AveragePrecision(activated_output_transform) """ def __init__(self, output_transform=lambda x: x): super(AveragePrecision, self).__init__(average_precision_compute_fn, output_transform=output_transform)

© Copyright 2024, PyTorch-Ignite Contributors. Last updated on 07/17/2024, 10:10:30 AM.

Built with Sphinx using a theme provided by Read the Docs.