guill
a5da4d0b3e
Fix error with ExecutionBlocker and OUTPUT_IS_LIST ( #4836 )
...
This change resolves an error when a node with OUTPUT_IS_LIST=(True,)
receives an ExecutionBlocker. I've also added a unit test for this case.
2024-09-08 09:48:47 -04:00
comfyanonymous
32a60a7bac
Support onetrainer text encoder Flux lora.
2024-09-08 09:31:41 -04:00
Jim Winkens
bb52934ba4
Fix import issue ( #4815 )
2024-09-07 05:28:32 -04:00
comfyanonymous
8aabd7c8c0
SaveLora node can now save "full diff" lora format.
...
This isn't actually a lora format and is saving the full diff of the
weights in a format that can be used in the lora loader nodes.
2024-09-07 03:21:02 -04:00
comfyanonymous
a09b29ca11
Add an option to the SaveLora node to store the bias diff.
2024-09-07 03:03:30 -04:00
comfyanonymous
9bfee68773
LoraSave node now supports generating text encoder loras.
...
text_encoder_diff should be connected to a CLIPMergeSubtract node.
model_diff and text_encoder_diff are optional inputs so you can create
model only loras, text encoder only loras or a lora that contains both.
2024-09-07 02:30:12 -04:00
comfyanonymous
ea77750759
Support a generic Comfy format for text encoder loras.
...
This is a format with keys like:
text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.v_proj.lora_up.weight
Instead of waiting for me to add support for specific lora formats you can
convert your text encoder loras to this format instead.
If you want to see an example save a text encoder lora with the SaveLora
node with the commit right after this one.
2024-09-07 02:20:39 -04:00
comfyanonymous
c27ebeb1c2
Fix onnx export not working on flux.
2024-09-06 03:21:52 -04:00
guill
0c7c98a965
Nodes using UNIQUE_ID as input are NOT_IDEMPOTENT ( #4793 )
...
As suggested by @ltdrdata, we can automatically consider nodes that take
the UNIQUE_ID hidden input to be NOT_IDEMPOTENT.
2024-09-05 19:33:02 -04:00
comfyanonymous
dc2eb75b85
Update stable release workflow to latest pytorch with cuda 12.4.
2024-09-05 19:21:52 -04:00
Chenlei Hu
fa34efe3bd
Update frontend to v1.2.47 ( #4798 )
...
* Update web content to release v1.2.47
* Update shortcut list
2024-09-05 18:56:01 -04:00
comfyanonymous
5cbaa9e07c
Mistoline flux controlnet support.
2024-09-05 00:05:17 -04:00
comfyanonymous
c7427375ee
Prioritize freeing partially offloaded models first.
2024-09-04 19:47:32 -04:00
comfyanonymous
22d1241a50
Add an experimental LoraSave node to extract model loras.
...
The model_diff input should be connected to the output of a
ModelMergeSubtract node.
2024-09-04 16:38:38 -04:00
Jedrzej Kosinski
f04229b84d
Add emb_patch support to UNetModel forward ( #4779 )
2024-09-04 14:35:15 -04:00
Silver
f067ad15d1
Make live preview size a configurable launch argument ( #4649 )
...
* Make live preview size a configurable launch argument
* Remove import from testing phase
* Update cli_args.py
2024-09-03 19:16:38 -04:00
comfyanonymous
483004dd1d
Support newer glora format.
2024-09-03 17:02:19 -04:00
comfyanonymous
00a5d08103
Lower fp8 lora memory usage.
2024-09-03 01:25:05 -04:00
comfyanonymous
d043997d30
Flux onetrainer lora.
2024-09-02 08:22:15 -04:00
Alex "mcmonkey" Goodwin
f1c2301697
fix typo in stale-issues ( #4735 )
2024-09-01 17:44:49 -04:00
comfyanonymous
8d31a6632f
Speed up inference on nvidia 10 series on Linux.
2024-09-01 17:29:31 -04:00
comfyanonymous
b643eae08b
Make minimum_inference_memory() depend on --reserve-vram
2024-09-01 01:18:34 -04:00
comfyanonymous
baa6b4dc36
Update manual install instructions.
2024-08-31 04:37:23 -04:00
Alex "mcmonkey" Goodwin
d4aeefc297
add github action to automatically handle stale user support issues ( #4683 )
...
* add github action to automatically handle stale user support issues
* improve stale message
* remove token part
2024-08-31 01:57:18 -04:00
comfyanonymous
587e7ca654
Remove github buttons.
2024-08-31 01:53:10 -04:00
Chenlei Hu
c90459eba0
Update ComfyUI_frontend to 1.2.40 ( #4691 )
...
* Update ComfyUI_frontend to 1.2.40
* Add files
2024-08-30 19:32:10 -04:00
Vedat Baday
04278afb10
feat: return import_failed from init_extra_nodes function ( #4694 )
2024-08-30 19:26:47 -04:00
comfyanonymous
935ae153e1
Cleanup.
2024-08-30 12:53:59 -04:00
Chenlei Hu
e91662e784
Get logs endpoint & system_stats additions ( #4690 )
...
* Add route for getting output logs
* Include ComfyUI version
* Move to own function
* Changed to memory logger
* Unify logger setup logic
* Fix get version git fallback
---------
Co-authored-by: pythongosssss <125205205+pythongosssss@users.noreply.github.com>
2024-08-30 12:46:37 -04:00
comfyanonymous
63fafaef45
Fix potential issue with hydit controlnets.
2024-08-30 04:58:41 -04:00
Alex "mcmonkey" Goodwin
ec28cd9136
swap legacy sdv15 link ( #4682 )
...
* swap legacy sdv15 link
* swap v15 ckpt examples to safetensors
* link the fp16 copy of the model by default
2024-08-29 19:48:48 -04:00
comfyanonymous
6eb5d64522
Fix glora lowvram issue.
2024-08-29 19:07:23 -04:00
comfyanonymous
10a79e9898
Implement model part of flux union controlnet.
2024-08-29 18:41:22 -04:00
comfyanonymous
ea3f39bd69
InstantX depth flux controlnet.
2024-08-29 02:14:19 -04:00
comfyanonymous
b33cd61070
InstantX canny controlnet.
2024-08-28 19:02:50 -04:00
Dr.Lt.Data
34eda0f853
fix: remove redundant useless loop ( #4656 )
...
fix: potential error of undefined variable
https://github.com/comfyanonymous/ComfyUI/discussions/4650
2024-08-28 17:46:30 -04:00
comfyanonymous
d31e226650
Unify RMSNorm code.
2024-08-28 16:56:38 -04:00
comfyanonymous
b79fd7d92c
ComfyUI supports more than just stable diffusion.
2024-08-28 16:12:24 -04:00
comfyanonymous
38c22e631a
Fix case where model was not properly unloaded in merging workflows.
2024-08-27 19:03:51 -04:00
Chenlei Hu
6bbdcd28ae
Support weight padding on diff weight patch ( #4576 )
2024-08-27 13:55:37 -04:00
comfyanonymous
ab130001a8
Do RMSNorm in native type.
2024-08-27 02:41:56 -04:00
Chenlei Hu
ca4b8f30e0
Cleanup empty dir if frontend zip download failed ( #4574 )
2024-08-27 02:07:25 -04:00
Robin Huang
70b84058c1
Add relative file path to the progress report. ( #4621 )
2024-08-27 02:06:12 -04:00
comfyanonymous
2ca8f6e23d
Make the stochastic fp8 rounding reproducible.
2024-08-26 15:12:06 -04:00
comfyanonymous
7985ff88b9
Use less memory in float8 lora patching by doing calculations in fp16.
2024-08-26 14:45:58 -04:00
comfyanonymous
c6812947e9
Fix potential memory leak.
2024-08-26 02:07:32 -04:00
comfyanonymous
9230f65823
Fix some controlnets OOMing when loading.
2024-08-25 05:54:29 -04:00
guill
6ab1e6fd4a
[Bug #4529 ] Fix graph partial validation failure ( #4588 )
...
Currently, if a graph partially fails validation (i.e. some outputs are
valid while others have links from missing nodes), the execution loop
could get an exception resulting in server lockup.
This isn't actually possible to reproduce via the default UI, but is a
potential issue for people using the API to construct invalid graphs.
2024-08-24 15:34:58 -04:00
comfyanonymous
07dcbc3a3e
Clarify how to use high quality previews.
2024-08-24 02:31:03 -04:00
comfyanonymous
8ae23d8e80
Fix onnx export.
2024-08-23 17:52:47 -04:00