• Docs >
  • Module code >
  • ignite.contrib.metrics.regression.mean_absolute_relative_error

Source code for ignite.contrib.metrics.regression.mean_absolute_relative_error

from __future__ import division
from ignite.exceptions import NotComputableError
from ignite.metrics import Metric
import torch

[docs]class MeanAbsoluteRelativeError(Metric): r""" Calculate Mean Absolute Relative Error: :math:`\text{MARE} = \frac{1}{n}\sum_{j=1}^n\frac{\left|A_j-P_j\right|}{\left|A_j\right|}`, where :math:`A_j` is the ground truth and :math:`P_j` is the predicted value. More details can be found in the reference `Botchkarev 2018`__. - `update` must receive output of the form `(y_pred, y)` __ https://arxiv.org/ftp/arxiv/papers/1809/1809.03006.pdf """
[docs] def reset(self): self._sum_of_absolute_relative_errors = 0.0 self._num_samples = 0
[docs] def update(self, output): y_pred, y = output if (y == 0).any(): raise NotComputableError('The ground truth has 0') absolute_error = torch.abs(y_pred - y.view_as(y_pred)) / torch.abs(y.view_as(y_pred)) self._sum_of_absolute_relative_errors += torch.sum(absolute_error).item() self._num_samples += y.size()[0]
[docs] def compute(self): if self._num_samples == 0: raise NotComputableError('MeanAbsoluteRelativeError must have at least' 'one sample before it can be computed') return self._sum_of_absolute_relative_errors / self._num_samples

© Copyright 2022, PyTorch-Ignite Contributors. Last updated on 05/04/2022, 8:31:16 PM.

Built with Sphinx using a theme provided by Read the Docs.