## ReLU {#openvino_docs_ops_activation_ReLU_1}
**Versioned name**: *ReLU-1*
**Category**: *Activation function*
**Short description**: ReLU element-wise activation function. ([Reference](http://caffe.berkeleyvision.org/tutorial/layers/relu.html))
**Detailed description**: [Reference](https://github.com/Kulbear/deep-learning-nano-foundation/wiki/ReLU-and-Softmax-Activation-Functions#rectified-linear-units)
**Attributes**: *ReLU* operation has no attributes.
**Mathematical Formulation**
For each element from the input tensor calculates corresponding
element in the output tensor with the following formula:
\f[
Y_{i}^{( l )} = max(0, Y_{i}^{( l - 1 )})
\f]
**Inputs**:
* **1**: Multidimensional input tensor *x* of any supported numeric type. **Required.**
**Outputs**:
* **1**: Result of ReLU function applied to the input tensor *x*. Tensor with shape and type matching the input tensor.
**Example**
```xml
25656
```