diff options
author | arch-fan <juanserracines@gmail.com> | 2023-06-09 19:59:27 +0000 |
---|---|---|
committer | arch-fan <juanserracines@gmail.com> | 2023-06-09 19:59:27 +0000 |
commit | 5576a7232207ea4df7c27f8ccd040d1697434532 (patch) | |
tree | ff2c7ea7a2264fe8a53defaf2766d1b76d43f2ce | |
parent | baf6946e06249c5af9851c60171692c44ef633e0 (diff) | |
download | stable-diffusion-webui-gfx803-5576a7232207ea4df7c27f8ccd040d1697434532.tar.gz stable-diffusion-webui-gfx803-5576a7232207ea4df7c27f8ccd040d1697434532.tar.bz2 stable-diffusion-webui-gfx803-5576a7232207ea4df7c27f8ccd040d1697434532.zip |
fixed typos
-rw-r--r-- | modules/shared.py | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/modules/shared.py b/modules/shared.py index 271a062d..e19f1576 100644 --- a/modules/shared.py +++ b/modules/shared.py @@ -416,9 +416,9 @@ options_templates.update(options_section(('sd', "Stable Diffusion"), { "enable_emphasis": OptionInfo(True, "Enable emphasis").info("use (text) to make model pay more attention to text and [text] to make it pay less attention"),
"enable_batch_seeds": OptionInfo(True, "Make K-diffusion samplers produce same images in a batch as when making a single image"),
"comma_padding_backtrack": OptionInfo(20, "Prompt word wrap length limit", gr.Slider, {"minimum": 0, "maximum": 74, "step": 1}).info("in tokens - for texts shorter than specified, if they don't fit into 75 token limit, move them to the next 75 token chunk"),
- "CLIP_stop_at_last_layers": OptionInfo(1, "Clip skip", gr.Slider, {"minimum": 1, "maximum": 12, "step": 1}).link("wiki", "https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Features#clip-skip").info("ignore last layers of CLIP nrtwork; 1 ignores none, 2 ignores one layer"),
+ "CLIP_stop_at_last_layers": OptionInfo(1, "Clip skip", gr.Slider, {"minimum": 1, "maximum": 12, "step": 1}).link("wiki", "https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Features#clip-skip").info("ignore last layers of CLIP network; 1 ignores none, 2 ignores one layer"),
"upcast_attn": OptionInfo(False, "Upcast cross attention layer to float32"),
- "randn_source": OptionInfo("GPU", "Random number generator source.", gr.Radio, {"choices": ["GPU", "CPU"]}).info("changes seeds drastically; use CPU to produce the same picture across different vidocard vendors"),
+ "randn_source": OptionInfo("GPU", "Random number generator source.", gr.Radio, {"choices": ["GPU", "CPU"]}).info("changes seeds drastically; use CPU to produce the same picture across different videocard vendors"),
}))
options_templates.update(options_section(('optimizations', "Optimizations"), {
|