diff options
author | AUTOMATIC1111 <16777216c@gmail.com> | 2023-06-27 06:19:55 +0000 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-06-27 06:19:55 +0000 |
commit | d4f9250c5aef0fd3afb6ed8a6bc3515fa2fcf635 (patch) | |
tree | 975b2cb80392d884b53df69370b5f4ed661ddfb3 | |
parent | 24129368f1b732be25ef486edb2cf5a6ace66737 (diff) | |
parent | d8071647760a2213aaf33a533addb4d84ba86816 (diff) | |
download | stable-diffusion-webui-gfx803-d4f9250c5aef0fd3afb6ed8a6bc3515fa2fcf635.tar.gz stable-diffusion-webui-gfx803-d4f9250c5aef0fd3afb6ed8a6bc3515fa2fcf635.tar.bz2 stable-diffusion-webui-gfx803-d4f9250c5aef0fd3afb6ed8a6bc3515fa2fcf635.zip |
Merge pull request #11201 from akx/ruff-upg
Upgrade Ruff to 0.0.272
-rw-r--r-- | .github/workflows/on_pull_request.yaml | 2 | ||||
-rw-r--r-- | modules/textual_inversion/logging.py | 48 |
2 files changed, 45 insertions, 5 deletions
diff --git a/.github/workflows/on_pull_request.yaml b/.github/workflows/on_pull_request.yaml index 7b7219fd..8ebf5918 100644 --- a/.github/workflows/on_pull_request.yaml +++ b/.github/workflows/on_pull_request.yaml @@ -18,7 +18,7 @@ jobs: # not to have GHA download an (at the time of writing) 4 GB cache # of PyTorch and other dependencies. - name: Install Ruff - run: pip install ruff==0.0.265 + run: pip install ruff==0.0.272 - name: Run Ruff run: ruff . lint-js: diff --git a/modules/textual_inversion/logging.py b/modules/textual_inversion/logging.py index 734a4b6f..45823eb1 100644 --- a/modules/textual_inversion/logging.py +++ b/modules/textual_inversion/logging.py @@ -2,11 +2,51 @@ import datetime import json
import os
-saved_params_shared = {"model_name", "model_hash", "initial_step", "num_of_dataset_images", "learn_rate", "batch_size", "clip_grad_mode", "clip_grad_value", "gradient_step", "data_root", "log_directory", "training_width", "training_height", "steps", "create_image_every", "template_file", "gradient_step", "latent_sampling_method"}
-saved_params_ti = {"embedding_name", "num_vectors_per_token", "save_embedding_every", "save_image_with_stored_embedding"}
-saved_params_hypernet = {"hypernetwork_name", "layer_structure", "activation_func", "weight_init", "add_layer_norm", "use_dropout", "save_hypernetwork_every"}
+saved_params_shared = {
+ "batch_size",
+ "clip_grad_mode",
+ "clip_grad_value",
+ "create_image_every",
+ "data_root",
+ "gradient_step",
+ "initial_step",
+ "latent_sampling_method",
+ "learn_rate",
+ "log_directory",
+ "model_hash",
+ "model_name",
+ "num_of_dataset_images",
+ "steps",
+ "template_file",
+ "training_height",
+ "training_width",
+}
+saved_params_ti = {
+ "embedding_name",
+ "num_vectors_per_token",
+ "save_embedding_every",
+ "save_image_with_stored_embedding",
+}
+saved_params_hypernet = {
+ "activation_func",
+ "add_layer_norm",
+ "hypernetwork_name",
+ "layer_structure",
+ "save_hypernetwork_every",
+ "use_dropout",
+ "weight_init",
+}
saved_params_all = saved_params_shared | saved_params_ti | saved_params_hypernet
-saved_params_previews = {"preview_prompt", "preview_negative_prompt", "preview_steps", "preview_sampler_index", "preview_cfg_scale", "preview_seed", "preview_width", "preview_height"}
+saved_params_previews = {
+ "preview_cfg_scale",
+ "preview_height",
+ "preview_negative_prompt",
+ "preview_prompt",
+ "preview_sampler_index",
+ "preview_seed",
+ "preview_steps",
+ "preview_width",
+}
def save_settings_to_file(log_directory, all_params):
|