Muhammed Yusuf
7891d13329
Added label for autoQueueCheckbox. ( #1295 )
...
* Added label for autoQueueCheckbox.
* Menu gets behind of some custom nodes.
* Edited extraOptions.
Options divided in to different divs to manage them with ease.
2023-09-02 02:58:23 -04:00
comfyanonymous
7931ff0fd9
Support SDXL inpaint models.
2023-09-01 15:22:52 -04:00
comfyanonymous
c335fdf200
Merge branch 'pixelass-patch-1' of https://github.com/pixelass/ComfyUI
2023-09-01 11:48:11 -04:00
comfyanonymous
43f2505389
Merge branch 'fix/widget-wonkyness' of https://github.com/M1kep/ComfyUI
2023-09-01 03:07:10 -04:00
comfyanonymous
0e3b641172
Remove xformers related print.
2023-09-01 02:12:03 -04:00
comfyanonymous
5c363a9d86
Fix controlnet bug.
2023-09-01 02:01:08 -04:00
Michael Poutre
69c5e6de85
fix(widgets): Add options object if not present when forceInput: true
2023-08-31 17:58:43 -07:00
Michael Poutre
9a7a52f8b5
refactor/fix: Treat forceInput widgets as standard widgets
2023-08-31 17:58:43 -07:00
comfyanonymous
cfe1c54de8
Fix controlnet issue.
2023-08-31 15:16:58 -04:00
comfyanonymous
57beace324
Fix VAEDecodeTiled minimum.
2023-08-31 14:26:16 -04:00
comfyanonymous
1c012d69af
It doesn't make sense for c_crossattn and c_concat to be lists.
2023-08-31 13:25:00 -04:00
comfyanonymous
5f101f4da1
Update litegraph with upstream: middle mouse dragging.
2023-08-31 02:39:34 -04:00
Ridan Vandenbergh
2cd3980199
Remove forced lowercase on embeddings endpoint
2023-08-30 20:48:55 +02:00
comfyanonymous
7e941f9f24
Clean up DiffusersLoader node.
2023-08-30 12:57:07 -04:00
Simon Lui
18617967e5
Fix error message in model_patcher.py
...
Found while tinkering.
2023-08-30 00:25:04 -07:00
comfyanonymous
fe4c07400c
Fix "Load Checkpoint with config" node.
2023-08-29 23:58:32 -04:00
comfyanonymous
d70b0bc43c
Use the GPU for the canny preprocessor when available.
2023-08-29 17:58:40 -04:00
comfyanonymous
81d9200e18
Add node to convert a specific colour in an image to a mask.
2023-08-29 17:55:42 -04:00
comfyanonymous
f2f5e5dcbb
Support SDXL t2i adapters with 3 channel input.
2023-08-29 16:44:57 -04:00
comfyanonymous
15adc3699f
Move beta_schedule to model_config and allow disabling unet creation.
2023-08-29 14:22:53 -04:00
comfyanonymous
968078b149
Merge branch 'feat/mute_bypass_nodes_in_group' of https://github.com/M1kep/ComfyUI
2023-08-29 11:33:40 -04:00
comfyanonymous
66c690e698
Merge branch 'preserve-pnginfo' of https://github.com/chrisgoringe/ComfyUI
2023-08-29 11:32:58 -04:00
comfyanonymous
bed116a1f9
Remove optimization that caused border.
2023-08-29 11:21:36 -04:00
Chris
18379dea36
check for text attr and save
2023-08-29 18:50:28 +10:00
Chris
edcff9ab8a
copy metadata into modified image
2023-08-29 18:50:28 +10:00
Michael Poutre
6944288aff
refactor(ui): Switch statement, and handle other modes in group actions
2023-08-29 00:24:31 -07:00
Michael Poutre
e30d546e38
feat(ui): Add node mode toggles to group context menu
2023-08-28 23:49:25 -07:00
comfyanonymous
8ddd081b09
Use the same units for tile size in VAEDecodeTiled and VAEEncodeTiled.
2023-08-29 01:51:35 -04:00
comfyanonymous
fbf375f161
Merge branch 'master' of https://github.com/bvhari/ComfyUI
2023-08-29 01:42:00 -04:00
comfyanonymous
65cae62c71
No need to check filename extensions to detect shuffle controlnet.
2023-08-28 16:49:06 -04:00
comfyanonymous
4e89b2c25a
Put clip vision outputs on the CPU.
2023-08-28 16:26:11 -04:00
comfyanonymous
a094b45c93
Load clipvision model to GPU for faster performance.
2023-08-28 15:29:27 -04:00
comfyanonymous
1300a1bb4c
Text encoder should initially load on the offload_device not the regular.
2023-08-28 15:08:45 -04:00
comfyanonymous
f92074b84f
Move ModelPatcher to model_patcher.py
2023-08-28 14:51:31 -04:00
BVH
d86b222fe9
Reduce min tile size for encode
2023-08-28 22:39:09 +05:30
comfyanonymous
4798cf5a62
Implement loras with norm keys.
2023-08-28 11:20:06 -04:00
BVH
9196588088
Make tile size in Tiled VAE encode/decode user configurable
2023-08-28 19:57:22 +05:30
Dr.Lt.Data
0faee1186f
support on prompt event handler ( #765 )
...
Co-authored-by: Lt.Dr.Data <lt.dr.data@gmail.com>
2023-08-28 00:52:22 -04:00
comfyanonymous
b8c7c770d3
Enable bf16-vae by default on ampere and up.
2023-08-27 23:06:19 -04:00
comfyanonymous
1c794a2161
Fallback to slice attention if xformers doesn't support the operation.
2023-08-27 22:24:42 -04:00
comfyanonymous
d935ba50c4
Make --bf16-vae work on torch 2.0
2023-08-27 21:33:53 -04:00
comfyanonymous
412596d325
Merge branch 'increase_client_max_size' of https://github.com/ramyma/ComfyUI
2023-08-27 13:12:39 -04:00
Dr.Lt.Data
d9f4922993
fix: cannot disable dynamicPrompts ( #1327 )
...
* fix: cannot disable dynamicPrompts
* indent fix
---------
Co-authored-by: Lt.Dr.Data <lt.dr.data@gmail.com>
2023-08-27 12:34:24 -04:00
ramyma
0b6cf7a558
Increase client_max_size to allow bigger request bodies
2023-08-26 19:48:20 +03:00
comfyanonymous
a57b0c797b
Fix lowvram model merging.
2023-08-26 11:52:07 -04:00
comfyanonymous
f72780a7e3
The new smart memory management makes this unnecessary.
2023-08-25 18:02:15 -04:00
comfyanonymous
c77f02e1c6
Move controlnet code to comfy/controlnet.py
2023-08-25 17:33:04 -04:00
comfyanonymous
15a7716fa6
Move lora code to comfy/lora.py
2023-08-25 17:11:51 -04:00
comfyanonymous
ec96f6d03a
Move text_projection to base clip model.
2023-08-24 23:43:48 -04:00
comfyanonymous
30eb92c3cb
Code cleanups.
2023-08-24 19:39:18 -04:00