Commit Graph

4574 Commits

Author SHA1 Message Date
doctorpangloss
3ee969f778 Test macOS 2025-07-16 15:48:22 -07:00
doctorpangloss
8fcccad6a2 fix linting error 2025-07-16 15:32:58 -07:00
doctorpangloss
3c5f477718 fix test 2025-07-16 15:26:19 -07:00
doctorpangloss
e455b0ad2e fix opentel 2025-07-16 15:19:19 -07:00
doctorpangloss
3ff24f0985 update dockerfile 2025-07-16 15:14:15 -07:00
doctorpangloss
6ecb9e71f6 this is not disabling the meterprovider successfully 2025-07-16 14:33:06 -07:00
doctorpangloss
5f77a57d33 fix amd dockerfile 2025-07-16 14:31:22 -07:00
doctorpangloss
5ec11ed92b add readme 2025-07-16 14:27:38 -07:00
doctorpangloss
8176078118 add configuration args types 2025-07-16 14:26:42 -07:00
doctorpangloss
94310e51e3 Merge branch 'master' of github.com:comfyanonymous/ComfyUI 2025-07-16 14:24:12 -07:00
doctorpangloss
d4d62500ac fix tests, disable opentel metrics provider 2025-07-16 14:21:50 -07:00
doctorpangloss
e67684e669 dockerfile updates 2025-07-16 14:11:05 -07:00
doctorpangloss
f23fe2e2cf updates to dockerfile and pyproject.yaml 2025-07-16 13:46:05 -07:00
comfyanonymous
491fafbd64
Silence clip tokenizer warning. (#8934) 2025-07-16 14:42:07 -04:00
Harel Cain
9bc2798f72
LTXV VAE decoder: switch default padding mode (#8930) 2025-07-16 13:54:38 -04:00
comfyanonymous
50afba747c
Add attempt to work around the safetensors mmap issue. (#8928) 2025-07-16 03:42:17 -04:00
Brandon Wallace
6b8062f414
Fix MaskComposite error when destination has 2 dimensions (#8915)
Fix code that is using the original `destination` input instead of the reshaped value.
2025-07-15 21:08:27 -04:00
doctorpangloss
c9f3de1b04 Update README to include ComfyUI Manager installation instructions. 2025-07-15 14:37:26 -07:00
doctorpangloss
25a0406b67 reduce metrics noise in tests 2025-07-15 14:08:46 -07:00
doctorpangloss
bf3345e083 Add support for ComfyUI Manager 2025-07-15 14:06:29 -07:00
doctorpangloss
4e91556820 Update README.md 2025-07-15 10:50:34 -07:00
doctorpangloss
80b967364c Fix tests 2025-07-15 10:48:11 -07:00
doctorpangloss
f1cfff6da5 Merge branch 'master' of github.com:comfyanonymous/ComfyUI 2025-07-15 10:21:19 -07:00
doctorpangloss
96b4e04315 packaging fixes
- enable user db
 - fix main_pre order everywhere
 - fix absolute to relative imports everywhere
 - async better supported
2025-07-15 10:19:33 -07:00
comfyanonymous
b1ae4126c3
Add action to detect windows line endings. (#8917) 2025-07-15 02:27:18 -04:00
Yoland Yan
9dabda19f0
Update nodes_gemini.py (#8912) 2025-07-14 20:59:35 -04:00
Yoland Yan
543c24108c
Fix wrong reference bug (#8910) 2025-07-14 20:45:55 -04:00
doctorpangloss
c086c5e005 fix pylint issues 2025-07-14 17:44:43 -07:00
doctorpangloss
499f9be5fa merging upstream with hiddenswitch
- deprecate our preview image fork of the frontend because upstream now has the needed functionality
 - merge the context executor from upstream into ours
2025-07-14 16:55:47 -07:00
doctorpangloss
bd6f28e3bd Move back to comfy_execution directory 2025-07-14 13:48:34 -07:00
doctorpangloss
04e411c32e Merge branch 'master' of github.com:comfyanonymous/ComfyUI 2025-07-14 13:45:09 -07:00
FeepingCreature
260a5ca5d9
Allow the prompt request to specify the prompt ID. (#8189)
This makes it easier to write asynchronous clients that submit requests, because they can store the task immediately.
Duplicate prompt IDs are rejected by the job queue.
2025-07-14 14:48:31 -04:00
ComfyUI Wiki
861c3bbb3d
Upate template to 0.1.36 (#8904) 2025-07-14 13:27:57 -04:00
comfyanonymous
9ca581c941
Remove windows line endings. (#8902) 2025-07-14 13:10:20 -04:00
comfyanonymous
4831e9c2c4
Refactor previous pr. (#8893) 2025-07-13 04:59:17 -04:00
Christian Byrne
480375f349
Remove auth tokens from history storage (#8889)
Remove auth_token_comfy_org and api_key_comfy_org from extra_data before
storing prompt history to prevent sensitive authentication tokens from
being persisted in the history endpoint response.
2025-07-13 04:46:27 -04:00
comfyanonymous
b40143984c
Add model detection error hint for lora. (#8880) 2025-07-12 03:49:26 -04:00
doctorpangloss
35899ba20e simplify composite function 2025-07-11 13:52:25 -07:00
doctorpangloss
a464dfa3db fix issue when trying to write to a file that already exists 2025-07-11 13:52:15 -07:00
doctorpangloss
f576f8124a record workflow when there's an error 2025-07-11 13:51:45 -07:00
chaObserv
b43916a134
Fix fresca's input and output (#8871) 2025-07-11 12:52:58 -04:00
JettHu
7bc7dd2aa2
Execute async node earlier (#8865) 2025-07-11 12:51:06 -04:00
comfyanonymous
938d3e8216
Remove windows line endings. (#8866) 2025-07-11 02:37:51 -04:00
Christian Byrne
8f05fb48ea
[fix] increase Kling API polling timeout to prevent user timeouts (#8860)
Extends polling duration from 10 minutes to ~68 minutes (256 attempts × 16 seconds) to accommodate longer Kling API operations that were frequently timing out for users.
2025-07-10 18:00:29 -04:00
comfyanonymous
b7ff5bd14d
Fix python3.9 (#8858) 2025-07-10 15:21:18 -04:00
guill
2b653e8c18
Support for async node functions (#8830)
* Support for async execution functions

This commit adds support for node execution functions defined as async. When
a node's execution function is defined as async, we can continue
executing other nodes while it is processing.

Standard uses of `await` should "just work", but people will still have
to be careful if they spawn actual threads. Because torch doesn't really
have async/await versions of functions, this won't particularly help
with most locally-executing nodes, but it does work for e.g. web
requests to other machines.

In addition to the execute function, the `VALIDATE_INPUTS` and
`check_lazy_status` functions can also be defined as async, though we'll
only resolve one node at a time right now for those.

* Add the execution model tests to CI

* Add a missing file

It looks like this got caught by .gitignore? There's probably a better
place to put it, but I'm not sure what that is.

* Add the websocket library for automated tests

* Add additional tests for async error cases

Also fixes one bug that was found when an async function throws an error
after being scheduled on a task.

* Add a feature flags message to reduce bandwidth

We now only send 1 preview message of the latest type the client can
support.

We'll add a console warning when the client fails to send a feature
flags message at some point in the future.

* Add async tests to CI

* Don't actually add new tests in this PR

Will do it in a separate PR

* Resolve unit test in GPU-less runner

* Just remove the tests that GHA can't handle

* Change line endings to UNIX-style

* Avoid loading model_management.py so early

Because model_management.py has a top-level `logging.info`, we have to
be careful not to import that file before we call `setup_logging`. If we
do, we end up having the default logging handler registered in addition
to our custom one.
2025-07-10 14:46:19 -04:00
comfyanonymous
1fd306824d
Add warning to catch torch import mistakes. (#8852) 2025-07-10 01:03:27 -04:00
Kohaku-Blueleaf
1205afc708
Better training loop implementation (#8820) 2025-07-09 11:41:22 -04:00
doctorpangloss
1338e8935f fix metrics exporter messing up health endpoint 2025-07-09 07:15:42 -07:00
doctorpangloss
b268296504 update upstream for flux fixes 2025-07-09 06:28:17 -07:00