Module: tf.keras.activations | TensorFlow v2.16.1 (original) (raw)
Module: tf.keras.activations
Stay organized with collections Save and categorize content based on your preferences.
DO NOT EDIT.
This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten.
Functions
deserialize(...): Return a Keras activation function via its config.
elu(...): Exponential Linear Unit.
exponential(...): Exponential activation function.
gelu(...): Gaussian error linear unit (GELU) activation function.
get(...): Retrieve a Keras activation function via an identifier.
hard_sigmoid(...): Hard sigmoid activation function.
hard_silu(...): Hard SiLU activation function, also known as Hard Swish.
hard_swish(...): Hard SiLU activation function, also known as Hard Swish.
leaky_relu(...): Leaky relu activation function.
linear(...): Linear activation function (pass-through).
log_softmax(...): Log-Softmax activation function.
mish(...): Mish activation function.
relu(...): Applies the rectified linear unit activation function.
relu6(...): Relu6 activation function.
selu(...): Scaled Exponential Linear Unit (SELU).
sigmoid(...): Sigmoid activation function.
silu(...): Swish (or Silu) activation function.
softmax(...): Softmax converts a vector of values to a probability distribution.
softplus(...): Softplus activation function.
softsign(...): Softsign activation function.
swish(...): Swish (or Silu) activation function.
tanh(...): Hyperbolic tangent activation function.