From ce6314bf5eb824000611deaaeff628150ac404b7 Mon Sep 17 00:00:00 2001 From: Hoel Bagard Date: Fri, 22 Jan 2021 12:48:33 +0900 Subject: [PATCH] Fixed bias --- layers.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/layers.py b/layers.py index 3bedb24..aaf219f 100644 --- a/layers.py +++ b/layers.py @@ -56,7 +56,7 @@ class Conv1d(Layer): super().__init__(activation, use_batch_norm) self.conv = nn.Conv1d(in_channels, out_channels, kernel_size, stride=stride, - bias=not Layer.USE_BATCH_NORM, **kwargs) + bias=not self.use_batch_norm, **kwargs) self.batch_norm = nn.BatchNorm1d( out_channels, momentum=Layer.BATCH_NORM_MOMENTUM, @@ -72,7 +72,7 @@ class Conv2d(Layer): super().__init__(activation, use_batch_norm) self.conv = nn.Conv2d(in_channels, out_channels, kernel_size, stride=stride, - bias=not Layer.USE_BATCH_NORM, **kwargs) + bias=not self.use_batch_norm, **kwargs) self.batch_norm = nn.BatchNorm2d( out_channels, momentum=Layer.BATCH_NORM_MOMENTUM, @@ -88,7 +88,7 @@ class Conv3d(Layer): super().__init__(activation, use_batch_norm) self.conv = nn.Conv3d(in_channels, out_channels, kernel_size, stride=stride, - bias=not Layer.USE_BATCH_NORM, **kwargs) + bias=not self.use_batch_norm, **kwargs) self.batch_norm = nn.BatchNorm3d( out_channels, momentum=Layer.BATCH_NORM_MOMENTUM, @@ -105,7 +105,7 @@ class Deconv2d(Layer): self.deconv = nn.ConvTranspose2d( in_channels, out_channels, kernel_size, stride=stride, - bias=not Layer.USE_BATCH_NORM, **kwargs) + bias=not self.use_batch_norm, **kwargs) self.batch_norm = nn.BatchNorm2d( out_channels, momentum=Layer.BATCH_NORM_MOMENTUM,