ReLU

Versioned name : ReLU-1

Category : Activation function

Short description : ReLU element-wise activation function. (Reference)

Detailed description : Reference

Attributes : ReLU operation has no attributes.

Mathematical Formulation

For each element from the input tensor calculates corresponding element in the output tensor with the following formula:

\[Y_{i}^{( l )} = max(0,\ Y_{i}^{( l - 1 )})\]

Inputs :

  • 1 : Multidimensional input tensor x of any supported numeric type. Required.

Outputs :

  • 1 : Result of ReLU function applied to the input tensor x. Tensor with shape and type matching the input tensor.

Example

<layer ... type="ReLU">
    <input>
        <port id="0">
            <dim>256</dim>
            <dim>56</dim>
        </port>
    </input>
    <output>
        <port id="1">
            <dim>256</dim>
            <dim>56</dim>
        </port>
    </output>
</layer>