@@ -39,8 +39,8 @@ _apply_scale_bias(x, scale, bias) = x .* scale .+ bias
3939
4040Shared code path for all built-in norm functions.
4141
42- `μ` and `σ²` should be calculated on the fly using [`NNlib. norm_stats`](@ref),
43- or extracted from an existing collection such as [`NNlib. RunningStats`](@ref).
42+ `μ` and `σ²` should be calculated on the fly using [`norm_stats`](@ref),
43+ or extracted from an existing collection such as [`RunningStats`](@ref).
4444`bias` and `scale` are consistent with cuDNN and Flux.Scale.
4545We opt for `scale` over `weight` to avoid confusion with dense layers.
4646If the size of the statistics and affine parameters differ,
@@ -64,7 +64,7 @@ Contains running mean and variance estimates for stateful norm functions.
6464If the parameters are mutable, they will be updated in-place.
6565Otherwise, they will be replaced wholesale.
6666
67- See also [`NNlib. update_running_stats!`](@ref).
67+ See also [`update_running_stats!`](@ref).
6868"""
6969mutable struct RunningStats{M <: AbstractArray , V <: AbstractArray , MT <: Real }
7070 mean:: M
@@ -114,10 +114,10 @@ end
114114 reduce_dims) where {N}
115115
116116Performs a moving average update for layers with tracked statistics.
117- `μ` and `σ²` are the sample mean and variance, most likely from [`NNlib. norm_stats`](@ref).
118- `reduce_dims` should also match the `dims` argument of [`NNlib. norm_stats`](@ref).
117+ `μ` and `σ²` are the sample mean and variance, most likely from [`norm_stats`](@ref).
118+ `reduce_dims` should also match the `dims` argument of [`norm_stats`](@ref).
119119
120- See also [`NNlib. RunningStats`](@ref).
120+ See also [`RunningStats`](@ref).
121121"""
122122function update_running_stats! (stats:: RunningStats , x, μ, σ², reduce_dims:: Dims )
123123 V = eltype (σ²)
@@ -153,7 +153,7 @@ Normalizes `x` along the first `S` dimensions.
153153
154154For an additional learned affine transform, provide a `S`-dimensional `scale` and `bias`.
155155
156- See also [`NNlib. batchnorm`](@ref), [`NNlib. instancenorm`](@ref), and [`NNlib. groupnorm`](@ref).
156+ See also [`batchnorm`](@ref), [`instancenorm`](@ref), and [`groupnorm`](@ref).
157157
158158# Examples
159159
@@ -190,14 +190,14 @@ Functional [Batch Normalization](https://arxiv.org/abs/1502.03167) operation.
190190Normalizes `x` along each ``D_1×...×D_{N-2}×1×D_N`` input slice,
191191where `N-1` is the "channel" (or "feature", for 2D inputs) dimension.
192192
193- Provide a [`NNlib. RunningStats`](@ref) to fix a estimated mean and variance.
193+ Provide a [`RunningStats`](@ref) to fix a estimated mean and variance.
194194`batchnorm` will renormalize the input using these statistics during inference,
195195and update them using batch-level statistics when training.
196196To override this behaviour, manually set a value for `training`.
197197
198198If specified, `scale` and `bias` will be applied as an additional learned affine transform.
199199
200- See also [`NNlib. layernorm`](@ref), [`NNlib. instancenorm`](@ref), and [`NNlib. groupnorm`](@ref).
200+ See also [`layernorm`](@ref), [`instancenorm`](@ref), and [`groupnorm`](@ref).
201201"""
202202function batchnorm (x:: AbstractArray{<:Any, N} ,
203203 running_stats:: Union{RunningStats, Nothing} = nothing ,
@@ -232,7 +232,7 @@ To override this behaviour, manually set a value for `training`.
232232
233233If specified, `scale` and `bias` will be applied as an additional learned affine transform.
234234
235- See also [`NNlib. layernorm`](@ref), [`NNlib. batchnorm`](@ref), and [`NNlib. groupnorm`](@ref).
235+ See also [`layernorm`](@ref), [`batchnorm`](@ref), and [`groupnorm`](@ref).
236236"""
237237function instancenorm (x:: AbstractArray{<:Any, N} ,
238238 running_stats:: Union{RunningStats, Nothing} = nothing ,
@@ -266,7 +266,7 @@ The number of channels must be an integer multiple of the number of groups.
266266
267267If specified, `scale` and `bias` will be applied as an additional learned affine transform.
268268
269- See also [`NNlib. layernorm`](@ref), [`NNlib. batchnorm`](@ref), and [`NNlib. instancenorm`](@ref).
269+ See also [`layernorm`](@ref), [`batchnorm`](@ref), and [`instancenorm`](@ref).
270270
271271# Examples
272272
0 commit comments