diff options
author | flamelaw <flamelaw.com3d2@gmail.com> | 2022-11-22 17:49:01 +0000 |
---|---|---|
committer | flamelaw <flamelaw.com3d2@gmail.com> | 2022-11-22 17:49:01 +0000 |
commit | 89d8ecff09b426ddc89eb5b432825f8f4c218051 (patch) | |
tree | 8e9670a9f985db140b3cde8c7987909f235ba1f6 | |
parent | 5b57f61ba47f8b11d19a5b46e7fb5a52458abae5 (diff) | |
download | stable-diffusion-webui-gfx803-89d8ecff09b426ddc89eb5b432825f8f4c218051.tar.gz stable-diffusion-webui-gfx803-89d8ecff09b426ddc89eb5b432825f8f4c218051.tar.bz2 stable-diffusion-webui-gfx803-89d8ecff09b426ddc89eb5b432825f8f4c218051.zip |
small fixes
-rw-r--r-- | modules/hypernetworks/hypernetwork.py | 6 | ||||
-rw-r--r-- | modules/textual_inversion/textual_inversion.py | 2 |
2 files changed, 4 insertions, 4 deletions
diff --git a/modules/hypernetworks/hypernetwork.py b/modules/hypernetworks/hypernetwork.py index 0128419b..4541af18 100644 --- a/modules/hypernetworks/hypernetwork.py +++ b/modules/hypernetworks/hypernetwork.py @@ -435,8 +435,8 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, gradient_step, optimizer_name = hypernetwork.optimizer_name
else:
print(f"Optimizer type {hypernetwork.optimizer_name} is not defined!")
- optimizer = torch.optim.AdamW(params=weights, lr=scheduler.learn_rate)
- optimizer_name = 'AdamW'
+ optimizer = torch.optim.AdamW(params=weights, lr=scheduler.learn_rate)
+ optimizer_name = 'AdamW'
if hypernetwork.optimizer_state_dict: # This line must be changed if Optimizer type can be different from saved optimizer.
try:
@@ -582,7 +582,7 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, gradient_step, shared.state.textinfo = f"""
<p>
Loss: {loss_step:.7f}<br/>
-Step: {hypernetwork.step}<br/>
+Step: {steps_done}<br/>
Last prompt: {html.escape(batch.cond_text[0])}<br/>
Last saved hypernetwork: {html.escape(last_saved_file)}<br/>
Last saved image: {html.escape(last_saved_image)}<br/>
diff --git a/modules/textual_inversion/textual_inversion.py b/modules/textual_inversion/textual_inversion.py index 3036e48a..fee08e33 100644 --- a/modules/textual_inversion/textual_inversion.py +++ b/modules/textual_inversion/textual_inversion.py @@ -436,7 +436,7 @@ def train_embedding(embedding_name, learn_rate, batch_size, gradient_step, data_ shared.state.textinfo = f"""
<p>
Loss: {loss_step:.7f}<br/>
-Step: {embedding.step}<br/>
+Step: {steps_done}<br/>
Last prompt: {html.escape(batch.cond_text[0])}<br/>
Last saved embedding: {html.escape(last_saved_file)}<br/>
Last saved image: {html.escape(last_saved_image)}<br/>
|