diff options
author | AUTOMATIC <16777216c@gmail.com> | 2022-10-08 19:22:15 +0300 |
---|---|---|
committer | AUTOMATIC <16777216c@gmail.com> | 2022-10-08 19:22:15 +0300 |
commit | 3061cdb7b610d4ba7f1ea695d9d6364b591e5bc7 (patch) | |
tree | 67ee49e2cc10f0ff154ec03470a0ca9296e4e2d2 /modules/shared.py | |
parent | f9c5da159245bb1e7603b3c8b9e0703bcb1c2ff5 (diff) |
add --force-enable-xformers option and also add messages to console regarding cross attention optimizations
Diffstat (limited to 'modules/shared.py')
-rw-r--r-- | modules/shared.py | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/modules/shared.py b/modules/shared.py index 02cb2722..8f941226 100644 --- a/modules/shared.py +++ b/modules/shared.py @@ -44,6 +44,7 @@ parser.add_argument("--scunet-models-path", type=str, help="Path to directory wi parser.add_argument("--swinir-models-path", type=str, help="Path to directory with SwinIR model file(s).", default=os.path.join(models_path, 'SwinIR'))
parser.add_argument("--ldsr-models-path", type=str, help="Path to directory with LDSR model file(s).", default=os.path.join(models_path, 'LDSR'))
parser.add_argument("--xformers", action='store_true', help="enable xformers for cross attention layers")
+parser.add_argument("--force-enable-xformers", action='store_true', help="enable xformers for cross attention layers regardless of whether the checking code thinks you can run it; do not make bug reports if this fails to work")
parser.add_argument("--opt-split-attention", action='store_true', help="force-enables cross-attention layer optimization. By default, it's on for torch.cuda and off for other torch devices.")
parser.add_argument("--disable-opt-split-attention", action='store_true', help="force-disables cross-attention layer optimization")
parser.add_argument("--opt-split-attention-v1", action='store_true', help="enable older version of split attention optimization that does not consume all the VRAM it can find")
|