HardSigmoid¶
Versioned name: HardSigmoid-1
Category: Activation function
Short description: HardSigmoid element-wise activation function.
Attributes: HardSigmoid operation has no attributes.
Mathematical Formulation
For each element from the input tensor calculates corresponding element in the output tensor with the following formula:
\[y = max(0,\ min(1,\ \alpha x + \beta))\]
where α corresponds to alpha
scalar input and β corresponds to beta
scalar input.
Inputs
1: An tensor of type T. Required.
2:
alpha
0D tensor (scalar) of type T. Required.3:
beta
0D tensor (scalar) of type T. Required.
Outputs
1: The result of the hard sigmoid operation. A tensor of type T.
Types
T: any floating-point type.
Examples
<layer ... type="HardSigmoid">
<input>
<port id="0">
<dim>256</dim>
<dim>56</dim>
</port>
<port id="1"/>
<port id="2"/>
</input>
<output>
<port id="3">
<dim>256</dim>
<dim>56</dim>
</port>
</output>
</layer>