Commit Graph

620 Commits

Author SHA1 Message Date
layerdiffusion d1d0ec46aa Maintain patching related
1. fix several problems related to layerdiffuse not unloaded
2. fix several problems related to Fooocus inpaint
3. Slightly speed up on-the-fly LoRAs by precomputing them to computation dtype
2024-08-30 15:18:21 -07:00
layerdiffusion df598c4df5 add "Diffusion in Low Bits" to info text 2024-08-22 23:23:14 -07:00
layerdiffusion 08f7487590 revise "send distilled_cfg_scale when generating hires conds" 2024-08-22 06:35:13 -07:00
DenOfEquity bfd7fb1d9f send distilled_cfg_scale when generating hires conds (#1403) 2024-08-22 06:33:48 -07:00
layerdiffusion 14ac95f908 fix 2024-08-20 01:37:01 -07:00
layerdiffusion d38e560e42 Implement some rethinking about LoRA system
1. Add an option to allow users to use UNet in fp8/gguf but lora in fp16.
2. All FP16 loras do not need patch. Others will only patch again when lora weight change.
3. FP8 unet + fp16 lora are available (somewhat only available) in Forge now. This also solves some “LoRA too subtle” problems.
4. Significantly speed up all gguf models (in Async mode) by using independent thread (CUDA stream) to compute and dequant at the same time, even when low-bit weights are already on GPU.
5. View “online lora” as a module similar to ControlLoRA so that it is moved to GPU together with model when sampling, achieving significant speedup and perfect low VRAM management simultaneously.
2024-08-19 04:31:59 -07:00
DenOfEquity 93bfd7f85b invalidate cond cache if distilled CFG changed (#1240)
* Update processing.py

add distilled_cfg_scale to params that invalidate cond cache

* Update ui.py

distilled CFG and CFG step size 0.1 (from 0.5)
2024-08-17 19:34:11 -07:00
layerdiffusion 0f266c48bd make clear_prompt_cache a function 2024-08-17 19:00:13 -07:00
lllyasviel 61f83dd610 support all flux models 2024-08-13 05:42:17 -07:00
layerdiffusion 19b41b9438 Add option to experiment with results from other impl
Setting -> Compatibility -> Try to reproduce the results from external software
2024-08-11 17:02:50 -07:00
lllyasviel cfa5242a75 forge 2.0.0
see also discussions
2024-08-10 19:24:19 -07:00
layerdiffusion 593455c4de global unload after env var change 2024-08-08 22:34:55 -07:00
layerdiffusion 20e1ba4a82 fix 2024-08-08 15:08:20 -07:00
lllyasviel 6921420b3f Load Model only when click Generate
#964
2024-08-08 14:51:13 -07:00
layerdiffusion a189f3e53e distilled cfg scale info text 2024-08-08 01:22:27 -07:00
layerdiffusion 572a3c3d8b distilled cfg scale 2024-08-08 00:57:40 -07:00
layerdiffusion 653be0a3ad comments 2024-08-08 00:20:15 -07:00
layerdiffusion 396d9f378d we do not need to waste 10 seconds on T5 when CFG=1 2024-08-08 00:19:16 -07:00
lllyasviel a6baf4a4b5 revise kernel
and add unused files
2024-08-07 16:51:24 -07:00
lllyasviel 71eaa5ca12 rework UI so that the toolbar is managed by Forge 2024-08-06 13:54:06 -07:00
layerdiffusion ae1d995d0d Finally removed model_hijack
finally
2024-08-05 21:05:25 -07:00
layerdiffusion 27641043ee remove dirty cond_stage_key 2024-08-05 11:39:20 -07:00
layerdiffusion d77582aa5a fix img2img inpaint model 2024-08-05 11:36:08 -07:00
layerdiffusion bccf9fb23a Free WebUI from its Prison
Congratulations WebUI. Say Hello to freedom.
2024-08-05 04:21:35 -07:00
layerdiffusion 0863765173 rework sd1.5 and sdxl from scratch 2024-08-05 03:08:17 -07:00
layerdiffusion 9679232d29 fix TI embedding info text
original webui is also broken but forge is fixed now
2024-08-04 18:49:35 -07:00
layerdiffusion a72154405e Text Processing Engine is Finished
100% reproduce all previous results, including TI embeddings, LoRAs in CLIP, emphasize settings, BREAK, timestep swap scheduling, AB mixture, advanced uncond, etc
Backend is 85% finished
2024-08-04 18:42:51 -07:00
layerdiffusion 8c087f920e rename files 2024-08-03 15:54:39 -07:00
layerdiffusion f6a0c69f0f fix ADetailer 2024-07-26 15:11:12 -07:00
layerdiffusion e26abf87ec Gradio 4 + WebUI 1.10 2024-07-26 12:02:46 -07:00
lllyasviel 10b5ca2541 vae already sliced in inner loop 2024-03-08 00:40:33 -08:00
Chengsong Zhang b9705c58f6 fix alphas cumprod (#478)
* fix alphas cumprod

* indentation
2024-03-04 01:03:32 -06:00
lllyasviel 95bcea72b1 Revert "fix alphas cumprod (#475)"
This reverts commit 72139b000c.
2024-03-03 22:40:56 -08:00
Chengsong Zhang 72139b000c fix alphas cumprod (#475) 2024-03-03 20:09:04 -06:00
lllyasviel 5166a723c2 bring back tiling
close #215
2024-02-25 21:29:35 -08:00
lllyasviel 16caff3d14 Revert "try solve #381"
This reverts commit 9d4c88912d.
2024-02-23 16:13:40 -08:00
lllyasviel 9d4c88912d try solve #381
caused by some potential historical webui problems
2024-02-23 16:01:40 -08:00
lllyasviel bde779a526 apply_token_merging 2024-02-23 15:43:27 -08:00
lllyasviel 95ddac3117 Merge upstream
upstream
2024-02-21 23:56:45 -08:00
Andray 33c8fe1221 avoid doble upscaling in inpaint 2024-02-19 16:57:49 +04:00
lllyasviel 30c8d742b3 Merge branch 'main' into upt 2024-02-11 16:42:36 -08:00
AUTOMATIC1111 e2b19900ec add infotext entry for emphasis; put emphasis into a separate file, add an option to parse but still ignore emphasis 2024-02-11 09:39:51 +03:00
Chenlei Hu 388ca351f4 Revert "Fix ruff linter (#137)" (#143)
This reverts commit 6b3ad64388.
2024-02-08 21:24:04 -05:00
Chenlei Hu 6b3ad64388 Fix ruff linter (#137)
* Fix ruff linter

* Remove unused imports

* Remove unused imports
2024-02-08 20:35:20 -05:00
lllyasviel 760f727eb9 use better context manager to fix potential problems 2024-02-08 01:51:18 -08:00
lllyasviel c11490d560 add_alphas_cumprod_modifier 2024-02-03 22:33:18 -08:00
lllyasviel 905a027fc1 Update processing.py 2024-02-02 23:01:03 -08:00
lllyasviel 1edd626f7c Update processing.py 2024-02-02 22:30:55 -08:00
lllyasviel 66e563fe03 inpaint done! 2024-01-30 15:13:26 -08:00
lllyasviel d04765cf3d Update processing.py 2024-01-27 22:47:04 -08:00