Applies the Exponential Linear Unit (ELU) function, element-wise, as described in the paper: Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs).
ELU is defined as:
alpha – the value for the ELU formulation. Default: 1.0
inplace – can optionally do the operation in-place. Default:
Input: , where means any number of dimensions.
Output: , same shape as the input.
>>> m = nn.ELU() >>> input = torch.randn(2) >>> output = m(input)