diff --git a/src/Transformations.jl b/src/Transformations.jl index c256f05..f6d8f8e 100644 --- a/src/Transformations.jl +++ b/src/Transformations.jl @@ -14,7 +14,7 @@ import LearnBase: transform, transform!, learn! # import StatsBase: logistic, logit import OnlineStats -import OnlineStats: weight, Weight, BoundedEqualWeight, Variances +import OnlineStats: weight, Weight, BoundedEqualWeight export input_node, @@ -173,7 +173,7 @@ include("linear.jl") include("affine.jl") include("misc.jl") include("layernorm.jl") -include("inputnorm.jl") +# include("inputnorm.jl") include("activations.jl") include("whiten.jl") include("chain.jl") diff --git a/src/inputnorm.jl b/src/inputnorm.jl index 4bbd551..f1fba20 100644 --- a/src/inputnorm.jl +++ b/src/inputnorm.jl @@ -2,6 +2,9 @@ #= Similar to Batch Normalization, except online and without the rescaling/skew y = (a .- μ) ./ σ + + TODO: This is currently broken because OnlineStats.Variances no longer + exists. =# type InputNorm{T,W<:Weight} <: Transformation n::Int # maps n --> n