layer_norm¶
- paddle.nn.functional. layer_norm ( x, normalized_shape, weight=None, bias=None, epsilon=1e-05, name=None ) [source]
-
nn.LayerNorm is recommended. For more information, please refer to LayerNorm .
- Parameters
-
x (Tensor) – Input Tensor. It’s data type should be bfloat16, float16, float32, float64.
normalized_shape (int|list|tuple) – Input shape from an expected input of size \([*, normalized_shape[0], normalized_shape[1], ..., normalized_shape[-1]]\). If it is a single integer, this module will normalize over the last dimension which is expected to be of that specific size.
weight (Tensor, optional) – The weight tensor of batch_norm. Default: None.
bias (Tensor, optional) – The bias tensor of batch_norm. Default: None.
epsilon (float, optional) – The small value added to the variance to prevent division by zero. Default: 1e-05.
name (str, optional) – Name for the LayerNorm, default is None. For more information, please refer to Name .
- Returns
-
None
Examples
import paddle x = paddle.rand((2, 2, 2, 3)) layer_norm_out = paddle.nn.functional.layer_norm(x, x.shape[1:]) print(layer_norm_out)