diff options
author | AUTOMATIC <16777216c@gmail.com> | 2023-05-10 08:37:18 +0000 |
---|---|---|
committer | AUTOMATIC <16777216c@gmail.com> | 2023-05-10 08:37:18 +0000 |
commit | a5121e7a0623db328a9462d340d389ed6737374a (patch) | |
tree | 2fa6051d457555ef2d61793af2756a44c0ea221c /modules/hypernetworks/hypernetwork.py | |
parent | 550256db1ce18778a9d56ff343d844c61b9f9b83 (diff) | |
download | stable-diffusion-webui-gfx803-a5121e7a0623db328a9462d340d389ed6737374a.tar.gz stable-diffusion-webui-gfx803-a5121e7a0623db328a9462d340d389ed6737374a.tar.bz2 stable-diffusion-webui-gfx803-a5121e7a0623db328a9462d340d389ed6737374a.zip |
fixes for B007
Diffstat (limited to 'modules/hypernetworks/hypernetwork.py')
-rw-r--r-- | modules/hypernetworks/hypernetwork.py | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/modules/hypernetworks/hypernetwork.py b/modules/hypernetworks/hypernetwork.py index 6ef0bfdf..38ef074f 100644 --- a/modules/hypernetworks/hypernetwork.py +++ b/modules/hypernetworks/hypernetwork.py @@ -177,34 +177,34 @@ class Hypernetwork: def weights(self):
res = []
- for k, layers in self.layers.items():
+ for layers in self.layers.values():
for layer in layers:
res += layer.parameters()
return res
def train(self, mode=True):
- for k, layers in self.layers.items():
+ for layers in self.layers.values():
for layer in layers:
layer.train(mode=mode)
for param in layer.parameters():
param.requires_grad = mode
def to(self, device):
- for k, layers in self.layers.items():
+ for layers in self.layers.values():
for layer in layers:
layer.to(device)
return self
def set_multiplier(self, multiplier):
- for k, layers in self.layers.items():
+ for layers in self.layers.values():
for layer in layers:
layer.multiplier = multiplier
return self
def eval(self):
- for k, layers in self.layers.items():
+ for layers in self.layers.values():
for layer in layers:
layer.eval()
for param in layer.parameters():
@@ -619,7 +619,7 @@ def train_hypernetwork(id_task, hypernetwork_name, learn_rate, batch_size, gradi try:
sd_hijack_checkpoint.add()
- for i in range((steps-initial_step) * gradient_step):
+ for _ in range((steps-initial_step) * gradient_step):
if scheduler.finished:
break
if shared.state.interrupted:
|