| View source on GitHub | 
Applies Alpha Dropout to the input.
tf.keras.layers.AlphaDropout(
    rate, noise_shape=None, seed=None, **kwargs
)
  Alpha Dropout is a Dropout that keeps mean and variance of inputs to their original values, in order to ensure the self-normalizing property even after this dropout. Alpha Dropout fits well to Scaled Exponential Linear Units by randomly setting activations to the negative saturation value.
| Args | |
|---|---|
| rate | float, drop probability (as with Dropout). The multiplicative noise will have standard deviationsqrt(rate / (1 - rate)). | 
| seed | Integer, optional random seed to enable deterministic behavior. | 
inputs: Input tensor (of any rank).training: Python boolean indicating whether the layer should behave in training mode (adding dropout) or in inference mode (doing nothing).Arbitrary. Use the keyword argument input_shape (tuple of integers, does not include the samples axis) when using this layer as the first layer in a model.
Same shape as input.
    © 2022 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 4.0.
Code samples licensed under the Apache 2.0 License.
    https://www.tensorflow.org/versions/r2.9/api_docs/python/tf/keras/layers/AlphaDropout