DO NOT EDIT.
This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten.
Functions
average_pool(...)
: Average pooling operation.
batch_normalization(...)
: Normalizes x
by mean
and variance
.
binary_crossentropy(...)
: Computes binary cross-entropy loss between target and output tensor.
categorical_crossentropy(...)
: Computes categorical cross-entropy loss between target and output tensor.
conv(...)
: General N-D convolution.
conv_transpose(...)
: General N-D convolution transpose.
ctc_decode(...)
: Decodes the output of a CTC model.
ctc_loss(...)
: CTC (Connectionist Temporal Classification) loss.
depthwise_conv(...)
: General N-D depthwise convolution.
elu(...)
: Exponential Linear Unit activation function.
gelu(...)
: Gaussian Error Linear Unit (GELU) activation function.
hard_sigmoid(...)
: Hard sigmoid activation function.
hard_silu(...)
: Hard SiLU activation function, also known as Hard Swish.
hard_swish(...)
: Hard SiLU activation function, also known as Hard Swish.
leaky_relu(...)
: Leaky version of a Rectified Linear Unit activation function.
log_sigmoid(...)
: Logarithm of the sigmoid activation function.
log_softmax(...)
: Log-softmax activation function.
max_pool(...)
: Max pooling operation.
moments(...)
: Calculates the mean and variance of x
.
multi_hot(...)
: Encodes integer labels as multi-hot vectors.
normalize(...)
: Normalizes x
over the specified axis.
one_hot(...)
: Converts integer tensor x
into a one-hot tensor.
psnr(...)
: Peak Signal-to-Noise Ratio (PSNR) function.
relu(...)
: Rectified linear unit activation function.
relu6(...)
: Rectified linear unit activation function with upper bound of 6.
selu(...)
: Scaled Exponential Linear Unit (SELU) activation function.
separable_conv(...)
: General N-D separable convolution.
sigmoid(...)
: Sigmoid activation function.
silu(...)
: Sigmoid Linear Unit (SiLU) activation function, also known as Swish.
softmax(...)
: Softmax activation function.
softplus(...)
: Softplus activation function.
softsign(...)
: Softsign activation function.
sparse_categorical_crossentropy(...)
: Computes sparse categorical cross-entropy loss.
swish(...)
: Sigmoid Linear Unit (SiLU) activation function, also known as Swish.