Module: tf.keras.activations | TensorFlow v2.16.1 (original) (raw)
DO NOT EDIT.
This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten.
Functions
deserialize(...): Return a Keras activation function via its config.
elu(...): Exponential Linear Unit.
exponential(...): Exponential activation function.
gelu(...): Gaussian error linear unit (GELU) activation function.
get(...): Retrieve a Keras activation function via an identifier.
hard_sigmoid(...): Hard sigmoid activation function.
hard_silu(...): Hard SiLU activation function, also known as Hard Swish.
hard_swish(...): Hard SiLU activation function, also known as Hard Swish.
leaky_relu(...): Leaky relu activation function.
linear(...): Linear activation function (pass-through).
log_softmax(...): Log-Softmax activation function.
mish(...): Mish activation function.
relu(...): Applies the rectified linear unit activation function.
relu6(...): Relu6 activation function.
selu(...): Scaled Exponential Linear Unit (SELU).
sigmoid(...): Sigmoid activation function.
silu(...): Swish (or Silu) activation function.
softmax(...): Softmax converts a vector of values to a probability distribution.
softplus(...): Softplus activation function.
softsign(...): Softsign activation function.
swish(...): Swish (or Silu) activation function.
tanh(...): Hyperbolic tangent activation function.