diff options
author | brkirch <brkirch@users.noreply.github.com> | 2023-01-06 06:01:51 +0000 |
---|---|---|
committer | brkirch <brkirch@users.noreply.github.com> | 2023-01-06 06:01:51 +0000 |
commit | b95a4c0ce5ab9c414e0494193bfff665f45e9e65 (patch) | |
tree | 64709ed9c735f1e89da0d0d5ddcad6a32005a5f8 /modules/shared.py | |
parent | b119815333026164f2bd7d1ca71f3e4f7a9afd0d (diff) | |
download | stable-diffusion-webui-gfx803-b95a4c0ce5ab9c414e0494193bfff665f45e9e65.tar.gz stable-diffusion-webui-gfx803-b95a4c0ce5ab9c414e0494193bfff665f45e9e65.tar.bz2 stable-diffusion-webui-gfx803-b95a4c0ce5ab9c414e0494193bfff665f45e9e65.zip |
Change sub-quad chunk threshold to use percentage
Diffstat (limited to 'modules/shared.py')
-rw-r--r-- | modules/shared.py | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/modules/shared.py b/modules/shared.py index cb1dc312..d7a81db1 100644 --- a/modules/shared.py +++ b/modules/shared.py @@ -59,7 +59,7 @@ parser.add_argument("--opt-split-attention", action='store_true', help="force-en parser.add_argument("--opt-sub-quad-attention", action='store_true', help="enable memory efficient sub-quadratic cross-attention layer optimization")
parser.add_argument("--sub-quad-q-chunk-size", type=int, help="query chunk size for the sub-quadratic cross-attention layer optimization to use", default=1024)
parser.add_argument("--sub-quad-kv-chunk-size", type=int, help="kv chunk size for the sub-quadratic cross-attention layer optimization to use", default=None)
-parser.add_argument("--sub-quad-chunk-threshold", type=int, help="the size threshold in bytes for the sub-quadratic cross-attention layer optimization to use chunking", default=None)
+parser.add_argument("--sub-quad-chunk-threshold", type=int, help="the percentage of VRAM threshold for the sub-quadratic cross-attention layer optimization to use chunking", default=None)
parser.add_argument("--opt-split-attention-invokeai", action='store_true', help="force-enables InvokeAI's cross-attention layer optimization. By default, it's on when cuda is unavailable.")
parser.add_argument("--opt-split-attention-v1", action='store_true', help="enable older version of split attention optimization that does not consume all the VRAM it can find")
parser.add_argument("--disable-opt-split-attention", action='store_true', help="force-disables cross-attention layer optimization")
|