diff options
author | d8ahazard <d8ahazard@gmail.com> | 2022-09-30 20:28:37 +0000 |
---|---|---|
committer | d8ahazard <d8ahazard@gmail.com> | 2022-09-30 20:28:37 +0000 |
commit | 8deae077004f0332ca607fc3a5d568b1a4705bec (patch) | |
tree | 1932930dd9eaeb89928a838055b9c26f431e7cd3 /modules/shared.py | |
parent | e82ea202997cbcd2ab72891cd075d9ba270eb67d (diff) | |
download | stable-diffusion-webui-gfx803-8deae077004f0332ca607fc3a5d568b1a4705bec.tar.gz stable-diffusion-webui-gfx803-8deae077004f0332ca607fc3a5d568b1a4705bec.tar.bz2 stable-diffusion-webui-gfx803-8deae077004f0332ca607fc3a5d568b1a4705bec.zip |
Add ScuNET DeNoiser/Upscaler
Q&D Implementation of ScuNET, thanks to our handy model loader. :P
https://github.com/cszn/SCUNet
Diffstat (limited to 'modules/shared.py')
-rw-r--r-- | modules/shared.py | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/modules/shared.py b/modules/shared.py index 8428c7a3..a48b995a 100644 --- a/modules/shared.py +++ b/modules/shared.py @@ -40,6 +40,7 @@ parser.add_argument("--gfpgan-models-path", type=str, help="Path to directory wi parser.add_argument("--esrgan-models-path", type=str, help="Path to directory with ESRGAN model file(s).", default=os.path.join(model_path, 'ESRGAN'))
parser.add_argument("--bsrgan-models-path", type=str, help="Path to directory with BSRGAN model file(s).", default=os.path.join(model_path, 'BSRGAN'))
parser.add_argument("--realesrgan-models-path", type=str, help="Path to directory with RealESRGAN model file(s).", default=os.path.join(model_path, 'RealESRGAN'))
+parser.add_argument("--scunet-models-path", type=str, help="Path to directory with ScuNET model file(s).", default=os.path.join(model_path, 'ScuNET'))
parser.add_argument("--swinir-models-path", type=str, help="Path to directory with SwinIR model file(s).", default=os.path.join(model_path, 'SwinIR'))
parser.add_argument("--ldsr-models-path", type=str, help="Path to directory with LDSR model file(s).", default=os.path.join(model_path, 'LDSR'))
parser.add_argument("--opt-split-attention", action='store_true', help="force-enables cross-attention layer optimization. By default, it's on for torch.cuda and off for other torch devices.")
|