diff options
author | AUTOMATIC <16777216c@gmail.com> | 2022-11-20 16:39:20 +0300 |
---|---|---|
committer | AUTOMATIC <16777216c@gmail.com> | 2022-11-20 16:39:20 +0300 |
commit | c81d440d876dfd2ab3560410f37442ef56fc6632 (patch) | |
tree | 338a1c12e7d6aa0cada63de74e31be2115c907d7 /modules/shared.py | |
parent | 47a44c7e421b98ca07e92dbf88769b04c9e28f86 (diff) |
moved deepdanbooru to pure pytorch implementation
Diffstat (limited to 'modules/shared.py')
-rw-r--r-- | modules/shared.py | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/modules/shared.py b/modules/shared.py index a4457305..c93ae2a3 100644 --- a/modules/shared.py +++ b/modules/shared.py @@ -55,7 +55,7 @@ parser.add_argument("--ldsr-models-path", type=str, help="Path to directory with parser.add_argument("--clip-models-path", type=str, help="Path to directory with CLIP model file(s).", default=None)
parser.add_argument("--xformers", action='store_true', help="enable xformers for cross attention layers")
parser.add_argument("--force-enable-xformers", action='store_true', help="enable xformers for cross attention layers regardless of whether the checking code thinks you can run it; do not make bug reports if this fails to work")
-parser.add_argument("--deepdanbooru", action='store_true', help="enable deepdanbooru interrogator")
+parser.add_argument("--deepdanbooru", action='store_true', help="does not do anything")
parser.add_argument("--opt-split-attention", action='store_true', help="force-enables Doggettx's cross-attention layer optimization. By default, it's on for torch cuda.")
parser.add_argument("--opt-split-attention-invokeai", action='store_true', help="force-enables InvokeAI's cross-attention layer optimization. By default, it's on when cuda is unavailable.")
parser.add_argument("--opt-split-attention-v1", action='store_true', help="enable older version of split attention optimization that does not consume all the VRAM it can find")
|