Files
openvino/docs/ops/activation/ReLU_1.md
Tatiana Savina b3ea6ceefa DOCS shift to rst - Opset R (#17159)
* ops to rst

* sphinx transition

* try html tag

* try comment

* try code directive

* try code directive

* try highlight

* try concole directive

* try line directive

* add highlight for code

* another directive

* introduce consoke directive

* add code format
2023-04-24 11:02:09 +02:00

1.3 KiB

ReLU

@sphinxdirective

Versioned name: ReLU-1

Category: Activation function

Short description: ReLU element-wise activation function. (Reference <http://caffe.berkeleyvision.org/tutorial/layers/relu.html>__).

Detailed description: Reference <https://github.com/Kulbear/deep-learning-nano-foundation/wiki/ReLU-and-Softmax-Activation-Functions#rectified-linear-units>__.

Attributes: ReLU operation has no attributes.

Mathematical Formulation

For each element from the input tensor calculates corresponding element in the output tensor with the following formula:

.. math::

Y_{i}^{( l )} = max(0,\ Y_{i}^{( l - 1 )})

Inputs:

  • 1: Multidimensional input tensor x of any supported numeric type. Required.

Outputs:

  • 1: Result of ReLU function applied to the input tensor x. Tensor with shape and type matching the input tensor.

Example

.. code-block:: cpp

<layer ... type="ReLU">
    <input>
        <port id="0">
            <dim>256</dim>
            <dim>56</dim>
        </port>
    </input>
    <output>
        <port id="1">
            <dim>256</dim>
            <dim>56</dim>
        </port>
    </output>
</layer>

@endsphinxdirective