@@ -22,8 +22,8 @@ func NewLayerNorm(name string, normalized_shape int, dtype Dtype, g *Graph) *Lay
2222}
2323
2424func (m * LayerNorm ) LayerNorm (x * Tensor ) * Tensor {
25- op := x . LayerNorm ( m . weight , m . bias )
26- return op
25+
26+ return nil
2727}
2828
2929// BatchNorm 批归一化
@@ -52,8 +52,7 @@ func NewBatchNorm(name string, num_features int, dtype Dtype, g *Graph) *BatchNo
5252}
5353
5454func (m * BatchNorm ) BatchNorm (x * Tensor ) * Tensor {
55- op := x .BatchNorm (m .weight , m .bias , m .running_mean , m .running_var )
56- return op
55+ return nil
5756}
5857
5958// InstanceNorm 实例归一化
@@ -78,8 +77,7 @@ func NewInstanceNorm(name string, num_features int, dtype Dtype, g *Graph) *Inst
7877}
7978
8079func (m * InstanceNorm ) InstanceNorm (x * Tensor ) * Tensor {
81- op := x .InstanceNorm (m .weight , m .bias )
82- return op
80+ return nil
8381}
8482
8583// GroupNorm 组归一化
@@ -110,8 +108,7 @@ func NewGroupNorm(name string, num_groups, num_channels int, dtype Dtype, g *Gra
110108}
111109
112110func (m * GroupNorm ) GroupNorm (x * Tensor ) * Tensor {
113- op := x .GroupNorm (m .weight , m .bias , m .num_groups )
114- return op
111+ return nil
115112}
116113
117114// RMSNorm Root Mean Square Layer Normalization
@@ -137,6 +134,5 @@ func NewRMSNorm(name string, normalized_shape int, dtype Dtype, g *Graph) *RMSNo
137134}
138135
139136func (m * RMSNorm ) RMSNorm (x * Tensor ) * Tensor {
140- op := x .RMSNorm (m .weight , m .eps )
141- return op
137+ return nil
142138}
0 commit comments