ignite.contrib.metrics#
Contribution module of metrics
- class ignite.contrib.metrics.AveragePrecision(output_transform=<function AveragePrecision.<lambda>>)[source]#
Computes Average Precision accumulating predictions and the ground-truth during an epoch and applying sklearn.metrics.average_precision_score .
- Parameters
output_transform (callable, optional) – a callable that is used to transform the
Engine
’s process_function’s output into the form expected by the metric. This can be useful if, for example, you have a multi-output model and you want to compute the metric with respect to one of the outputs.
AveragePrecision expects y to be comprised of 0’s and 1’s. y_pred must either be probability estimates or confidence values. To apply an activation to y_pred, use output_transform as shown below:
def activated_output_transform(output): y_pred, y = output y_pred = torch.softmax(y_pred, dim=1) return y_pred, y avg_precision = AveragePrecision(activated_output_transform)
- class ignite.contrib.metrics.GpuInfo[source]#
Provides GPU information: a) used memory percentage, b) gpu utilization percentage values as Metric on each iterations.
Note
In case if gpu utilization reports “N/A” on a given GPU, corresponding metric value is not set.
Examples
# Default GPU measurements GpuInfo().attach(trainer, name='gpu') # metric names are 'gpu:X mem(%)', 'gpu:X util(%)' # Logging with TQDM ProgressBar(persist=True).attach(trainer, metric_names=['gpu:0 mem(%)', 'gpu:0 util(%)']) # Progress bar will looks like # Epoch [2/10]: [12/24] 50%|█████ , gpu:0 mem(%)=79, gpu:0 util(%)=59 [00:17<1:23] # Logging with Tensorboard tb_logger.attach(trainer, log_handler=OutputHandler(tag="training", metric_names='all'), event_name=Events.ITERATION_COMPLETED)
- attach(engine, name='gpu', event_name=Events.ITERATION_COMPLETED)[source]#
Attaches current metric to provided engine. On the end of engine’s run, engine.state.metrics dictionary will contain computed metric’s value under provided name.
- Parameters
engine (Engine) – the engine to which the metric must be attached
name (str) – the name of the metric to attach
usage (str or MetricUsage, optional) – the usage of the metric. Valid string values should be ‘EpochWise.usage_name’ (default) or ‘BatchWise.usage_name’.
Example:
metric = ... metric.attach(engine, "mymetric") assert "mymetric" in engine.run(data).metrics assert metric.is_attached(engine)
Example with usage:
metric = ... metric.attach(engine, "mymetric", usage=BatchWise.usage_name) assert "mymetric" in engine.run(data).metrics assert metric.is_attached(engine, usage=BatchWise.usage_name)
- completed(engine, name)[source]#
Helper method to compute metric’s value and put into the engine. It is automatically attached to the engine with
attach()
.- Parameters
engine (Engine) – the engine to which the metric must be attached
- compute()[source]#
Computes the metric based on it’s accumulated state.
By default, this is called at the end of each epoch.
- Returns
- the actual quantity of interest. However, if a
Mapping
is returned, it will be (shallow) flattened into engine.state.metrics when
completed()
is called.
- the actual quantity of interest. However, if a
- Return type
Any
- Raises
NotComputableError – raised when the metric cannot be computed.
- class ignite.contrib.metrics.PrecisionRecallCurve(output_transform=<function PrecisionRecallCurve.<lambda>>)[source]#
Compute precision-recall pairs for different probability thresholds for binary classification task by accumulating predictions and the ground-truth during an epoch and applying sklearn.metrics.precision_recall_curve .
- Parameters
output_transform (callable, optional) – a callable that is used to transform the
Engine
’s process_function’s output into the form expected by the metric. This can be useful if, for example, you have a multi-output model and you want to compute the metric with respect to one of the outputs.
PrecisionRecallCurve expects y to be comprised of 0’s and 1’s. y_pred must either be probability estimates or confidence values. To apply an activation to y_pred, use output_transform as shown below:
def activated_output_transform(output): y_pred, y = output y_pred = torch.sigmoid(y_pred) return y_pred, y roc_auc = PrecisionRecallCurve(activated_output_transform)
- class ignite.contrib.metrics.ROC_AUC(output_transform=<function ROC_AUC.<lambda>>)[source]#
Computes Area Under the Receiver Operating Characteristic Curve (ROC AUC) accumulating predictions and the ground-truth during an epoch and applying sklearn.metrics.roc_auc_score .
- Parameters
output_transform (callable, optional) – a callable that is used to transform the
Engine
’s process_function’s output into the form expected by the metric. This can be useful if, for example, you have a multi-output model and you want to compute the metric with respect to one of the outputs.
ROC_AUC expects y to be comprised of 0’s and 1’s. y_pred must either be probability estimates or confidence values. To apply an activation to y_pred, use output_transform as shown below:
def activated_output_transform(output): y_pred, y = output y_pred = torch.sigmoid(y_pred) return y_pred, y roc_auc = ROC_AUC(activated_output_transform)
- class ignite.contrib.metrics.RocCurve(output_transform=<function RocCurve.<lambda>>)[source]#
Compute Receiver operating characteristic (ROC) for binary classification task by accumulating predictions and the ground-truth during an epoch and applying sklearn.metrics.roc_curve .
- Parameters
output_transform (callable, optional) – a callable that is used to transform the
Engine
’s process_function’s output into the form expected by the metric. This can be useful if, for example, you have a multi-output model and you want to compute the metric with respect to one of the outputs.
RocCurve expects y to be comprised of 0’s and 1’s. y_pred must either be probability estimates or confidence values. To apply an activation to y_pred, use output_transform as shown below:
def activated_output_transform(output): y_pred, y = output y_pred = torch.sigmoid(y_pred) return y_pred, y roc_auc = RocCurve(activated_output_transform)
Regression metrics#
Module ignite.contrib.metrics.regression
provides implementations of
metrics useful for regression tasks. Definitions of metrics are based on Botchkarev 2018, page 30 “Appendix 2. Metrics mathematical definitions”.
Complete list of metrics:
- class ignite.contrib.metrics.regression.CanberraMetric(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Canberra Metric.
where, is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.FractionalAbsoluteError(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Fractional Absolute Error.
where, is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.FractionalBias(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Fractional Bias:
,
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.GeometricMeanAbsoluteError(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Geometric Mean Absolute Error.
where, is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.GeometricMeanRelativeAbsoluteError(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Geometric Mean Relative Absolute Error:
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.ManhattanDistance(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Manhattan Distance:
,
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.MaximumAbsoluteError(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Maximum Absolute Error:
,
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.MeanAbsoluteRelativeError(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculate Mean Absolute Relative Error:
,
where is the ground truth and is the predicted value.
More details can be found in the reference Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.MeanError(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Mean Error:
,
where is the ground truth and is the predicted value.
More details can be found in the reference Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.MeanNormalizedBias(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Mean Normalized Bias:
,
where is the ground truth and is the predicted value.
More details can be found in the reference Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).
- class ignite.contrib.metrics.regression.MedianAbsoluteError(output_transform=<function MedianAbsoluteError.<lambda>>)[source]#
Calculates the Median Absolute Error:
,
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1) and of type float32.
Warning
Current implementation stores all input data (output and target) in as tensors before computing a metric. This can potentially lead to a memory error if the input data is larger than available RAM.
- class ignite.contrib.metrics.regression.MedianAbsolutePercentageError(output_transform=<function MedianAbsolutePercentageError.<lambda>>)[source]#
Calculates the Median Absolute Percentage Error:
,
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1) and of type float32.
Warning
Current implementation stores all input data (output and target) in as tensors before computing a metric. This can potentially lead to a memory error if the input data is larger than available RAM.
- class ignite.contrib.metrics.regression.MedianRelativeAbsoluteError(output_transform=<function MedianRelativeAbsoluteError.<lambda>>)[source]#
Calculates the Median Relative Absolute Error:
,
where is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1) and of type float32.
Warning
Current implementation stores all input data (output and target) in as tensors before computing a metric. This can potentially lead to a memory error if the input data is larger than available RAM.
- class ignite.contrib.metrics.regression.R2Score(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the R-Squared, the coefficient of determination:
,
where is the ground truth, is the predicted value and is the mean of the ground truth.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1) and of type float32.
- class ignite.contrib.metrics.regression.WaveHedgesDistance(output_transform=<function Metric.<lambda>>, device=None)[source]#
Calculates the Wave Hedges Distance.
, where, is the ground truth and is the predicted value.
More details can be found in Botchkarev 2018.
update must receive output of the form (y_pred, y) or {‘y_pred’: y_pred, ‘y’: y}.
y and y_pred must be of same shape (N, ) or (N, 1).