Commit Graph

200 Commits (10bbbed218458b8a899aac2140ec738d8d716f05)

Author SHA1 Message Date
Vladimir Mandic d25b020f61 update 2023-06-02 12:29:21 -04:00
Disty0 d6b3504f4d Compile with IPEX Optimize 2023-06-02 13:04:50 +03:00
Vladimir Mandic 0ccda9bc8b jumbo patch 2023-05-17 14:15:55 -04:00
Vladimir Mandic 1921504e64 enable dynamo compile 2023-05-12 15:58:00 -04:00
Disty0 8171d57c36 Remove unnecessary IPEX imports 2023-05-04 02:34:34 +03:00
Vladimir Mandic cb4cff3929 redesign logging 2023-05-02 13:57:16 -04:00
Disty0 68fc95b2e1 Merge remote-tracking branch 'upstream/master' 2023-04-30 18:28:22 +03:00
Disty0 de8d0bef9f More patches and Import IPEX after Torch 2023-04-30 18:19:37 +03:00
Vladimir Mandic 682330b172 new command line parser 2023-04-30 10:54:59 -04:00
Vladimir Mandic 917ecad43c add dynamo options 2023-04-30 09:31:38 -04:00
Vladimir Mandic 57204b3d70 disable xformers/sdp if cannot be used 2023-04-21 11:32:19 -04:00
Vladimir Mandic 0e7144186d jump patch 2023-04-20 11:20:27 -04:00
Vladimir Mandic 86783d164c unload xformers when not used 2023-04-18 07:47:41 -04:00
Vladimir Mandic fd51bb90d0 enable quick launch 2023-04-15 11:51:58 -04:00
Vladimir Mandic ed8819b8fc lycoris, strong linting, model keyword, circular imports 2023-04-15 10:28:31 -04:00
Vladimir Mandic 81b8294e93 switch cmdflags to settings 2023-04-12 10:40:11 -04:00
Vladimir Mandic 99602e7454 add option to disable sdp 2023-03-27 10:19:42 -04:00
Vladimir Mandic f275e43eb9 update torch load and external repos 2023-03-20 10:10:56 -04:00
Vladimir Mandic f6679fcc77 add global exception handler 2023-03-17 10:08:07 -04:00
Pam 8d7fa2f67c sdp_attnblock_forward hijack 2023-03-10 22:48:41 +05:00
Pam 0981dea948 sdp refactoring 2023-03-10 12:58:10 +05:00
Pam 37acba2633 argument to disable memory efficient for sdp 2023-03-10 12:19:36 +05:00
Pam fec0a89511 scaled dot product attention 2023-03-07 00:33:13 +05:00
AUTOMATIC1111 dfb3b8f398
Merge branch 'master' into weighted-learning 2023-02-19 12:41:29 +03:00
Shondoit c4bfd20f31 Hijack to add weighted_forward to model: return loss * weight map 2023-02-15 10:03:59 +01:00
brkirch 2016733814 Apply hijacks in ddpm_edit for upcast sampling
To avoid import errors, ddpm_edit hijacks are done after an instruct pix2pix model is loaded.
2023-02-07 22:53:45 -05:00
AUTOMATIC1111 fecb990deb
Merge pull request #7309 from brkirch/fix-embeddings
Fix embeddings, upscalers, and refactor `--upcast-sampling`
2023-01-28 18:44:36 +03:00
AUTOMATIC d04e3e921e automatically detect v-parameterization for SD2 checkpoints 2023-01-28 15:24:41 +03:00
brkirch ada17dbd7c Refactor conditional casting, fix upscalers 2023-01-28 04:16:25 -05:00
brkirch c4b9b07db6 Fix embeddings dtype mismatch 2023-01-26 09:00:15 -05:00
AUTOMATIC 6073456c83 write a comment for fix_checkpoint function 2023-01-19 20:39:10 +03:00
AUTOMATIC 924e222004 add option to show/hide warnings
removed hiding warnings from LDSR
fixed/reworked few places that produced warnings
2023-01-18 23:04:24 +03:00
AUTOMATIC 085427de0e make it possible for extensions/scripts to add their own embedding directories 2023-01-08 09:37:33 +03:00
AUTOMATIC1111 c295e4a244
Merge pull request #6055 from brkirch/sub-quad_attn_opt
Add Birch-san's sub-quadratic attention implementation
2023-01-07 12:26:55 +03:00
AUTOMATIC 79e39fae61 CLIP hijack rework 2023-01-07 01:46:13 +03:00
brkirch 5deb2a19cc Allow Doggettx's cross attention opt without CUDA 2023-01-06 01:33:15 -05:00
brkirch 3bfe2bb549 Merge remote-tracking branch 'upstream/master' into sub-quad_attn_opt 2023-01-06 00:15:22 -05:00
brkirch f6ab5a39d7 Merge branch 'AUTOMATIC1111:master' into sub-quad_attn_opt 2023-01-06 00:14:20 -05:00
brkirch d782a95967 Add Birch-san's sub-quadratic attention implementation 2023-01-06 00:14:13 -05:00
Vladimir Mandic 21ee77db31
add cross-attention info 2023-01-04 08:04:38 -05:00
AUTOMATIC f34c734172 alt-diffusion integration 2022-12-31 18:06:35 +03:00
AUTOMATIC 3f401cdb64 Merge remote-tracking branch 'baai-open-internal/master' into alt-diffusion 2022-12-31 13:02:28 +03:00
AUTOMATIC 505ec7e4d9 cleanup some unneeded imports for hijack files 2022-12-10 09:17:39 +03:00
AUTOMATIC 7dbfd8a7d8 do not replace entire unet for the resolution hack 2022-12-10 09:14:45 +03:00
AUTOMATIC1111 2641d1b83b
Merge pull request #4978 from aliencaocao/support_any_resolution
Patch UNet Forward to support resolutions that are not multiples of 64
2022-12-10 08:45:41 +03:00
zhaohu xing 5dcc22606d add hash and fix undo hijack bug
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-12-06 16:04:50 +08:00
Zac Liu a25dfebeed
Merge pull request #3 from 920232796/master
fix device support for mps
update the support for SD2.0
2022-12-06 09:17:57 +08:00
Zac Liu 3ebf977a6e
Merge branch 'AUTOMATIC1111:master' into master 2022-12-06 09:16:15 +08:00
zhaohu xing 4929503258 fix bugs
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-12-06 09:03:55 +08:00
AUTOMATIC 0d21624cee move #5216 to the extension 2022-12-03 18:16:19 +03:00
AUTOMATIC 89e1df013b Merge remote-tracking branch 'wywywywy/autoencoder-hijack' 2022-12-03 18:08:10 +03:00
AUTOMATIC1111 a2feaa95fc
Merge pull request #5194 from brkirch/autocast-and-mps-randn-fixes
Use devices.autocast() and fix MPS randn issues
2022-12-03 09:58:08 +03:00
SmirkingFace da698ca92e Fixed AttributeError where openaimodel is not found 2022-12-02 13:47:02 +01:00
zhaohu xing 52cc83d36b fix bugs
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-11-30 14:56:12 +08:00
zhaohu xing 0831ab476c
Merge branch 'master' into master 2022-11-30 10:13:17 +08:00
wywywywy 36c3613d16
Add autoencoder to sd_hijack 2022-11-29 17:40:02 +00:00
zhaohu xing 75c4511e6b add AltDiffusion to webui
Signed-off-by: zhaohu xing <920232796@qq.com>
2022-11-29 10:28:41 +08:00
brkirch 98ca437edf Refactor and instead check if mps is being used, not availability 2022-11-28 21:18:51 -05:00
AUTOMATIC b48b7999c8 Merge remote-tracking branch 'flamelaw/master' 2022-11-27 12:19:59 +03:00
Billy Cao 349f0461ec
Merge branch 'master' into support_any_resolution 2022-11-27 12:39:31 +08:00
AUTOMATIC 64c7b7975c restore hypernetworks to seemingly working state 2022-11-26 16:45:57 +03:00
AUTOMATIC ce6911158b Add support Stable Diffusion 2.0 2022-11-26 16:10:46 +03:00
Billy Cao adb6cb7619 Patch UNet Forward to support resolutions that are not multiples of 64
Also modifed the UI to no longer step in 64
2022-11-23 18:11:24 +08:00
flamelaw bd68e35de3 Gradient accumulation, autocast fix, new latent sampling method, etc 2022-11-20 12:35:26 +09:00
killfrenzy96 17e4432820 cleanly undo circular hijack #4818 2022-11-18 21:22:55 +11:00
AUTOMATIC c62d17aee3 use the new devices.has_mps() function in register_buffer for DDIM/PLMS fix for OSX 2022-11-12 10:00:22 +03:00
AUTOMATIC 7ba3923d5b move DDIM/PLMS fix for OSX out of the file with inpainting code. 2022-11-11 18:20:18 +03:00
Jairo Correa af758e97fa Unload sd_model before loading the other 2022-11-01 04:01:49 -03:00
AUTOMATIC 2b91251637 removed aesthetic gradients as built-in
added support for extensions
2022-10-22 12:23:58 +03:00
AUTOMATIC 9286fe53de make aestetic embedding ciompatible with prompts longer than 75 tokens 2022-10-21 16:38:06 +03:00
AUTOMATIC 7d6b388d71 Merge branch 'ae' 2022-10-21 13:35:01 +03:00
C43H66N12O12S2 73b5dbf72a Update sd_hijack.py 2022-10-18 11:53:04 +03:00
C43H66N12O12S2 786ed49922 use legacy attnblock 2022-10-18 11:53:04 +03:00
MalumaDev 9324cdaa31 ui fix, re organization of the code 2022-10-16 17:53:56 +02:00
MalumaDev e4f8b5f00d ui fix 2022-10-16 10:28:21 +02:00
MalumaDev 523140d780 ui fix 2022-10-16 10:23:30 +02:00
MalumaDev b694bba39a Merge remote-tracking branch 'origin/test_resolve_conflicts' into test_resolve_conflicts 2022-10-16 00:24:05 +02:00
MalumaDev 9325c85f78 fixed dropbox update 2022-10-16 00:23:47 +02:00
MalumaDev 97ceaa23d0
Merge branch 'master' into test_resolve_conflicts 2022-10-16 00:06:36 +02:00
C43H66N12O12S2 529afbf4d7 Update sd_hijack.py 2022-10-15 20:25:27 +03:00
MalumaDev 37d7ffb415 fix to tokens lenght, addend embs generator, add new features to edit the embedding before the generation using text 2022-10-15 15:59:37 +02:00
MalumaDev bb57f30c2d init 2022-10-14 10:56:41 +02:00
AUTOMATIC 429442f4a6 fix iterator bug for #2295 2022-10-12 13:38:03 +03:00
hentailord85ez 80f3cf2bb2 Account when lines are mismatched 2022-10-12 11:38:41 +03:00
brkirch 98fd5cde72 Add check for psutil 2022-10-11 17:24:00 +03:00
brkirch c0484f1b98 Add cross-attention optimization from InvokeAI
* Add cross-attention optimization from InvokeAI (~30% speed improvement on MPS)
* Add command line option for it
* Make it default when CUDA is unavailable
2022-10-11 17:24:00 +03:00
AUTOMATIC 873efeed49 rename hypernetwork dir to hypernetworks to prevent clash with an old filename that people who use zip instead of git clone will have 2022-10-11 15:51:30 +03:00
AUTOMATIC 5de806184f Merge branch 'master' into hypernetwork-training 2022-10-11 11:14:36 +03:00
hentailord85ez 5e2627a1a6
Comma backtrack padding (#2192)
Comma backtrack padding
2022-10-11 09:55:28 +03:00
C43H66N12O12S2 623251ce2b allow pascal onwards 2022-10-10 19:54:07 +03:00
hentailord85ez d5c14365fd Add back in output hidden states parameter 2022-10-10 18:54:48 +03:00
hentailord85ez 460bbae587 Pad beginning of textual inversion embedding 2022-10-10 18:54:48 +03:00
hentailord85ez b340439586 Unlimited Token Works
Unlimited tokens actually work now. Works with textual inversion too. Replaces the previous not-so-much-working implementation.
2022-10-10 18:54:48 +03:00
Fampai 1824e9ee3a Removed unnecessary tmp variable 2022-10-09 22:31:23 +03:00
Fampai ad3ae44108 Updated code for legibility 2022-10-09 22:31:23 +03:00
Fampai e59c66c008 Optimized code for Ignoring last CLIP layers 2022-10-09 22:31:23 +03:00
Fampai 1371d7608b Added ability to ignore last n layers in FrozenCLIPEmbedder 2022-10-08 22:10:37 +03:00
AUTOMATIC 3061cdb7b6 add --force-enable-xformers option and also add messages to console regarding cross attention optimizations 2022-10-08 19:22:15 +03:00
C43H66N12O12S2 cc0258aea7 check for ampere without destroying the optimizations. again. 2022-10-08 17:54:16 +03:00
C43H66N12O12S2 017b6b8744 check for ampere 2022-10-08 17:54:16 +03:00