[FEATURE] Add scale to random initializer

This commit is contained in:
Jensun Ravichandran 2021-06-03 16:35:44 +02:00
parent ca8ac7a43b
commit 729b20e9ab

View File

@ -71,15 +71,16 @@ class ZerosInitializer(DimensionAwareInitializer):
class UniformInitializer(DimensionAwareInitializer): class UniformInitializer(DimensionAwareInitializer):
def __init__(self, dims, minimum=0.0, maximum=1.0): def __init__(self, dims, minimum=0.0, maximum=1.0, scale=1.0):
super().__init__(dims) super().__init__(dims)
self.minimum = minimum self.minimum = minimum
self.maximum = maximum self.maximum = maximum
self.scale = scale
def generate(self, length): def generate(self, length):
gen_dims = (length, ) + self.components_dims gen_dims = (length, ) + self.components_dims
return torch.ones(gen_dims).uniform_(self.minimum, self.maximum) return torch.ones(gen_dims).uniform_(self.minimum,
self.maximum) * self.scale
class DataAwareInitializer(ComponentsInitializer): class DataAwareInitializer(ComponentsInitializer):