MegEngine
This page contains details of all activation functions for MegEngine backend supported in Echo.
Mish
Applies the element-wise function:
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
Mish: A Self Regularized Non-Monotonic Activation Function
Swish
Allows the following element-wise functions:
Parameters:
eswish - Uses E-Swish activation function. Default:
False
.swish - Uses Swish activation function. Default:
False
.flatten - Uses Flatten T-Swish activation function. Default:
False
.beta - parameter used for E-Swish formulation. Default: 1.375
Note: When eswish, swish and flatten are False
, it initializes the SILU activation function by default.
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
References:
Searching for Activation Functions
E-swish: Adjusting Activations to Different Network Depths
Flatten-T Swish: a thresholded ReLU-Swish-like activation function for deep learning
Sigmoid-Weighted Linear Units for Neural Network Function Approximation in Reinforcement Learning
Aria2
Applies the element-wise function:
Parameters:
beta -is the exponential growth rate. Default: 0.5
alpha -is a hyper-parameter which has a two-fold effect; it reduces the curvature in 3rd quadrant as well as increases the curvature in first quadrant while lowering the value of activation. Default: 1.0
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
ELiSH
Allows the following element-wise functions:
Parameter:
hard - Uses Hard ELiSH activation function. Default:
False
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
The Quest for the Golden Activation Function
ISRU
Allows the following element-wise functions:
Parameters:
alpha - hyperparametercontrols the value to which an ISRLU saturates for negative inputs. Default: 1.0
isrlu - Uses ISRLU activation function. Default:
False
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
Improving Deep Learning by Inverse Square Root Linear Units (ISRLUs)
NLReLU
Applies the element-wise function:
Parameters:
beta - parameter used for NLReLU formulation. Default: 1.0
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
Natural-Logarithm-Rectified Activation Function in Convolutional Neural Networks
Soft Clipping
Applies the element-wise function:
Parameters:
alpha -hyper-parameter, which determines how close to linear the central region is and how sharply the linear region turns to the asymptotic values. Default: 0.5
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
Neural Network-Based Approach to Phase Space Integration
Soft Exponential
Applies the element-wise function:
Parameters:
alpha -trainable hyper-parameter which is initialized to zero by default. Default:
None
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
SQNL
Applies the element-wise function:
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
SQNL: A New Computationally Efficient Activation Function
SReLU
Applies the element-wise function:
Parameters:
in_features - Shape of the input. Datatype:
Tuple
parameters - ( ) parameters for manual initialization, Default:
None
. IfNone
is passed, parameters are initialized randomly.
Shape:
Input:wheremeans any number of additional dimensions
Output:, same shape as input
Reference:
Deep Learning with S-shaped Rectified Linear Activation Units
FReLU
Applies the element-wise function:
Parameter:
in_channels - Number of channels in the input tensor. Datatype:
Integer
Shape:
Input:whereindicates the number of channels.
Output:, same shape as input
Reference:
Last updated