Execution Model Inversion (#2666)
* Execution Model Inversion
This PR inverts the execution model -- from recursively calling nodes to
using a topological sort of the nodes. This change allows for
modification of the node graph during execution. This allows for two
major advantages:
1. The implementation of lazy evaluation in nodes. For example, if a
"Mix Images" node has a mix factor of exactly 0.0, the second image
input doesn't even need to be evaluated (and visa-versa if the mix
factor is 1.0).
2. Dynamic expansion of nodes. This allows for the creation of dynamic
"node groups". Specifically, custom nodes can return subgraphs that
replace the original node in the graph. This is an incredibly
powerful concept. Using this functionality, it was easy to
implement:
a. Components (a.k.a. node groups)
b. Flow control (i.e. while loops) via tail recursion
c. All-in-one nodes that replicate the WebUI functionality
d. and more
All of those were able to be implemented entirely via custom nodes,
so those features are *not* a part of this PR. (There are some
front-end changes that should occur before that functionality is
made widely available, particularly around variant sockets.)
The custom nodes associated with this PR can be found at:
https://github.com/BadCafeCode/execution-inversion-demo-comfyui
Note that some of them require that variant socket types ("*") be
enabled.
* Allow `input_info` to be of type `None`
* Handle errors (like OOM) more gracefully
* Add a command-line argument to enable variants
This allows the use of nodes that have sockets of type '*' without
applying a patch to the code.
* Fix an overly aggressive assertion.
This could happen when attempting to evaluate `IS_CHANGED` for a node
during the creation of the cache (in order to create the cache key).
* Fix Pyright warnings
* Add execution model unit tests
* Fix issue with unused literals
Behavior should now match the master branch with regard to undeclared
inputs. Undeclared inputs that are socket connections will be used while
undeclared inputs that are literals will be ignored.
* Make custom VALIDATE_INPUTS skip normal validation
Additionally, if `VALIDATE_INPUTS` takes an argument named `input_types`,
that variable will be a dictionary of the socket type of all incoming
connections. If that argument exists, normal socket type validation will
not occur. This removes the last hurdle for enabling variant types
entirely from custom nodes, so I've removed that command-line option.
I've added appropriate unit tests for these changes.
* Fix example in unit test
This wouldn't have caused any issues in the unit test, but it would have
bugged the UI if someone copy+pasted it into their own node pack.
* Use fstrings instead of '%' formatting syntax
* Use custom exception types.
* Display an error for dependency cycles
Previously, dependency cycles that were created during node expansion
would cause the application to quit (due to an uncaught exception). Now,
we'll throw a proper error to the UI. We also make an attempt to 'blame'
the most relevant node in the UI.
* Add docs on when ExecutionBlocker should be used
* Remove unused functionality
* Rename ExecutionResult.SLEEPING to PENDING
* Remove superfluous function parameter
* Pass None for uneval inputs instead of default
This applies to `VALIDATE_INPUTS`, `check_lazy_status`, and lazy values
in evaluation functions.
* Add a test for mixed node expansion
This test ensures that a node that returns a combination of expanded
subgraphs and literal values functions correctly.
* Raise exception for bad get_node calls.
* Minor refactor of IsChangedCache.get
* Refactor `map_node_over_list` function
* Fix ui output for duplicated nodes
* Add documentation on `check_lazy_status`
* Add file for execution model unit tests
* Clean up Javascript code as per review
* Improve documentation
Converted some comments to docstrings as per review
* Add a new unit test for mixed lazy results
This test validates that when an output list is fed to a lazy node, the
node will properly evaluate previous nodes that are needed by any inputs
to the lazy node.
No code in the execution model has been changed. The test already
passes.
* Allow kwargs in VALIDATE_INPUTS functions
When kwargs are used, validation is skipped for all inputs as if they
had been mentioned explicitly.
* List cached nodes in `execution_cached` message
This was previously just bugged in this PR.
2024-08-15 15:21:11 +00:00
|
|
|
import itertools
|
|
|
|
from typing import Sequence, Mapping
|
2024-08-15 13:37:30 +00:00
|
|
|
from comfy_execution.graph import DynamicPrompt
|
Execution Model Inversion (#2666)
* Execution Model Inversion
This PR inverts the execution model -- from recursively calling nodes to
using a topological sort of the nodes. This change allows for
modification of the node graph during execution. This allows for two
major advantages:
1. The implementation of lazy evaluation in nodes. For example, if a
"Mix Images" node has a mix factor of exactly 0.0, the second image
input doesn't even need to be evaluated (and visa-versa if the mix
factor is 1.0).
2. Dynamic expansion of nodes. This allows for the creation of dynamic
"node groups". Specifically, custom nodes can return subgraphs that
replace the original node in the graph. This is an incredibly
powerful concept. Using this functionality, it was easy to
implement:
a. Components (a.k.a. node groups)
b. Flow control (i.e. while loops) via tail recursion
c. All-in-one nodes that replicate the WebUI functionality
d. and more
All of those were able to be implemented entirely via custom nodes,
so those features are *not* a part of this PR. (There are some
front-end changes that should occur before that functionality is
made widely available, particularly around variant sockets.)
The custom nodes associated with this PR can be found at:
https://github.com/BadCafeCode/execution-inversion-demo-comfyui
Note that some of them require that variant socket types ("*") be
enabled.
* Allow `input_info` to be of type `None`
* Handle errors (like OOM) more gracefully
* Add a command-line argument to enable variants
This allows the use of nodes that have sockets of type '*' without
applying a patch to the code.
* Fix an overly aggressive assertion.
This could happen when attempting to evaluate `IS_CHANGED` for a node
during the creation of the cache (in order to create the cache key).
* Fix Pyright warnings
* Add execution model unit tests
* Fix issue with unused literals
Behavior should now match the master branch with regard to undeclared
inputs. Undeclared inputs that are socket connections will be used while
undeclared inputs that are literals will be ignored.
* Make custom VALIDATE_INPUTS skip normal validation
Additionally, if `VALIDATE_INPUTS` takes an argument named `input_types`,
that variable will be a dictionary of the socket type of all incoming
connections. If that argument exists, normal socket type validation will
not occur. This removes the last hurdle for enabling variant types
entirely from custom nodes, so I've removed that command-line option.
I've added appropriate unit tests for these changes.
* Fix example in unit test
This wouldn't have caused any issues in the unit test, but it would have
bugged the UI if someone copy+pasted it into their own node pack.
* Use fstrings instead of '%' formatting syntax
* Use custom exception types.
* Display an error for dependency cycles
Previously, dependency cycles that were created during node expansion
would cause the application to quit (due to an uncaught exception). Now,
we'll throw a proper error to the UI. We also make an attempt to 'blame'
the most relevant node in the UI.
* Add docs on when ExecutionBlocker should be used
* Remove unused functionality
* Rename ExecutionResult.SLEEPING to PENDING
* Remove superfluous function parameter
* Pass None for uneval inputs instead of default
This applies to `VALIDATE_INPUTS`, `check_lazy_status`, and lazy values
in evaluation functions.
* Add a test for mixed node expansion
This test ensures that a node that returns a combination of expanded
subgraphs and literal values functions correctly.
* Raise exception for bad get_node calls.
* Minor refactor of IsChangedCache.get
* Refactor `map_node_over_list` function
* Fix ui output for duplicated nodes
* Add documentation on `check_lazy_status`
* Add file for execution model unit tests
* Clean up Javascript code as per review
* Improve documentation
Converted some comments to docstrings as per review
* Add a new unit test for mixed lazy results
This test validates that when an output list is fed to a lazy node, the
node will properly evaluate previous nodes that are needed by any inputs
to the lazy node.
No code in the execution model has been changed. The test already
passes.
* Allow kwargs in VALIDATE_INPUTS functions
When kwargs are used, validation is skipped for all inputs as if they
had been mentioned explicitly.
* List cached nodes in `execution_cached` message
This was previously just bugged in this PR.
2024-08-15 15:21:11 +00:00
|
|
|
|
|
|
|
import nodes
|
|
|
|
|
2024-08-15 13:37:30 +00:00
|
|
|
from comfy_execution.graph_utils import is_link
|
Execution Model Inversion (#2666)
* Execution Model Inversion
This PR inverts the execution model -- from recursively calling nodes to
using a topological sort of the nodes. This change allows for
modification of the node graph during execution. This allows for two
major advantages:
1. The implementation of lazy evaluation in nodes. For example, if a
"Mix Images" node has a mix factor of exactly 0.0, the second image
input doesn't even need to be evaluated (and visa-versa if the mix
factor is 1.0).
2. Dynamic expansion of nodes. This allows for the creation of dynamic
"node groups". Specifically, custom nodes can return subgraphs that
replace the original node in the graph. This is an incredibly
powerful concept. Using this functionality, it was easy to
implement:
a. Components (a.k.a. node groups)
b. Flow control (i.e. while loops) via tail recursion
c. All-in-one nodes that replicate the WebUI functionality
d. and more
All of those were able to be implemented entirely via custom nodes,
so those features are *not* a part of this PR. (There are some
front-end changes that should occur before that functionality is
made widely available, particularly around variant sockets.)
The custom nodes associated with this PR can be found at:
https://github.com/BadCafeCode/execution-inversion-demo-comfyui
Note that some of them require that variant socket types ("*") be
enabled.
* Allow `input_info` to be of type `None`
* Handle errors (like OOM) more gracefully
* Add a command-line argument to enable variants
This allows the use of nodes that have sockets of type '*' without
applying a patch to the code.
* Fix an overly aggressive assertion.
This could happen when attempting to evaluate `IS_CHANGED` for a node
during the creation of the cache (in order to create the cache key).
* Fix Pyright warnings
* Add execution model unit tests
* Fix issue with unused literals
Behavior should now match the master branch with regard to undeclared
inputs. Undeclared inputs that are socket connections will be used while
undeclared inputs that are literals will be ignored.
* Make custom VALIDATE_INPUTS skip normal validation
Additionally, if `VALIDATE_INPUTS` takes an argument named `input_types`,
that variable will be a dictionary of the socket type of all incoming
connections. If that argument exists, normal socket type validation will
not occur. This removes the last hurdle for enabling variant types
entirely from custom nodes, so I've removed that command-line option.
I've added appropriate unit tests for these changes.
* Fix example in unit test
This wouldn't have caused any issues in the unit test, but it would have
bugged the UI if someone copy+pasted it into their own node pack.
* Use fstrings instead of '%' formatting syntax
* Use custom exception types.
* Display an error for dependency cycles
Previously, dependency cycles that were created during node expansion
would cause the application to quit (due to an uncaught exception). Now,
we'll throw a proper error to the UI. We also make an attempt to 'blame'
the most relevant node in the UI.
* Add docs on when ExecutionBlocker should be used
* Remove unused functionality
* Rename ExecutionResult.SLEEPING to PENDING
* Remove superfluous function parameter
* Pass None for uneval inputs instead of default
This applies to `VALIDATE_INPUTS`, `check_lazy_status`, and lazy values
in evaluation functions.
* Add a test for mixed node expansion
This test ensures that a node that returns a combination of expanded
subgraphs and literal values functions correctly.
* Raise exception for bad get_node calls.
* Minor refactor of IsChangedCache.get
* Refactor `map_node_over_list` function
* Fix ui output for duplicated nodes
* Add documentation on `check_lazy_status`
* Add file for execution model unit tests
* Clean up Javascript code as per review
* Improve documentation
Converted some comments to docstrings as per review
* Add a new unit test for mixed lazy results
This test validates that when an output list is fed to a lazy node, the
node will properly evaluate previous nodes that are needed by any inputs
to the lazy node.
No code in the execution model has been changed. The test already
passes.
* Allow kwargs in VALIDATE_INPUTS functions
When kwargs are used, validation is skipped for all inputs as if they
had been mentioned explicitly.
* List cached nodes in `execution_cached` message
This was previously just bugged in this PR.
2024-08-15 15:21:11 +00:00
|
|
|
|
|
|
|
class CacheKeySet:
|
|
|
|
def __init__(self, dynprompt, node_ids, is_changed_cache):
|
|
|
|
self.keys = {}
|
|
|
|
self.subcache_keys = {}
|
|
|
|
|
|
|
|
def add_keys(self, node_ids):
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def all_node_ids(self):
|
|
|
|
return set(self.keys.keys())
|
|
|
|
|
|
|
|
def get_used_keys(self):
|
|
|
|
return self.keys.values()
|
|
|
|
|
|
|
|
def get_used_subcache_keys(self):
|
|
|
|
return self.subcache_keys.values()
|
|
|
|
|
|
|
|
def get_data_key(self, node_id):
|
|
|
|
return self.keys.get(node_id, None)
|
|
|
|
|
|
|
|
def get_subcache_key(self, node_id):
|
|
|
|
return self.subcache_keys.get(node_id, None)
|
|
|
|
|
|
|
|
class Unhashable:
|
|
|
|
def __init__(self):
|
|
|
|
self.value = float("NaN")
|
|
|
|
|
|
|
|
def to_hashable(obj):
|
|
|
|
# So that we don't infinitely recurse since frozenset and tuples
|
|
|
|
# are Sequences.
|
|
|
|
if isinstance(obj, (int, float, str, bool, type(None))):
|
|
|
|
return obj
|
|
|
|
elif isinstance(obj, Mapping):
|
|
|
|
return frozenset([(to_hashable(k), to_hashable(v)) for k, v in sorted(obj.items())])
|
|
|
|
elif isinstance(obj, Sequence):
|
|
|
|
return frozenset(zip(itertools.count(), [to_hashable(i) for i in obj]))
|
|
|
|
else:
|
|
|
|
# TODO - Support other objects like tensors?
|
|
|
|
return Unhashable()
|
|
|
|
|
|
|
|
class CacheKeySetID(CacheKeySet):
|
|
|
|
def __init__(self, dynprompt, node_ids, is_changed_cache):
|
|
|
|
super().__init__(dynprompt, node_ids, is_changed_cache)
|
|
|
|
self.dynprompt = dynprompt
|
|
|
|
self.add_keys(node_ids)
|
|
|
|
|
|
|
|
def add_keys(self, node_ids):
|
|
|
|
for node_id in node_ids:
|
|
|
|
if node_id in self.keys:
|
|
|
|
continue
|
|
|
|
node = self.dynprompt.get_node(node_id)
|
|
|
|
self.keys[node_id] = (node_id, node["class_type"])
|
|
|
|
self.subcache_keys[node_id] = (node_id, node["class_type"])
|
|
|
|
|
|
|
|
class CacheKeySetInputSignature(CacheKeySet):
|
|
|
|
def __init__(self, dynprompt, node_ids, is_changed_cache):
|
|
|
|
super().__init__(dynprompt, node_ids, is_changed_cache)
|
|
|
|
self.dynprompt = dynprompt
|
|
|
|
self.is_changed_cache = is_changed_cache
|
|
|
|
self.add_keys(node_ids)
|
|
|
|
|
|
|
|
def include_node_id_in_input(self) -> bool:
|
|
|
|
return False
|
|
|
|
|
|
|
|
def add_keys(self, node_ids):
|
|
|
|
for node_id in node_ids:
|
|
|
|
if node_id in self.keys:
|
|
|
|
continue
|
|
|
|
node = self.dynprompt.get_node(node_id)
|
|
|
|
self.keys[node_id] = self.get_node_signature(self.dynprompt, node_id)
|
|
|
|
self.subcache_keys[node_id] = (node_id, node["class_type"])
|
|
|
|
|
|
|
|
def get_node_signature(self, dynprompt, node_id):
|
|
|
|
signature = []
|
|
|
|
ancestors, order_mapping = self.get_ordered_ancestry(dynprompt, node_id)
|
|
|
|
signature.append(self.get_immediate_node_signature(dynprompt, node_id, order_mapping))
|
|
|
|
for ancestor_id in ancestors:
|
|
|
|
signature.append(self.get_immediate_node_signature(dynprompt, ancestor_id, order_mapping))
|
|
|
|
return to_hashable(signature)
|
|
|
|
|
|
|
|
def get_immediate_node_signature(self, dynprompt, node_id, ancestor_order_mapping):
|
|
|
|
node = dynprompt.get_node(node_id)
|
|
|
|
class_type = node["class_type"]
|
|
|
|
class_def = nodes.NODE_CLASS_MAPPINGS[class_type]
|
|
|
|
signature = [class_type, self.is_changed_cache.get(node_id)]
|
|
|
|
if self.include_node_id_in_input() or (hasattr(class_def, "NOT_IDEMPOTENT") and class_def.NOT_IDEMPOTENT):
|
|
|
|
signature.append(node_id)
|
|
|
|
inputs = node["inputs"]
|
|
|
|
for key in sorted(inputs.keys()):
|
|
|
|
if is_link(inputs[key]):
|
|
|
|
(ancestor_id, ancestor_socket) = inputs[key]
|
|
|
|
ancestor_index = ancestor_order_mapping[ancestor_id]
|
|
|
|
signature.append((key,("ANCESTOR", ancestor_index, ancestor_socket)))
|
|
|
|
else:
|
|
|
|
signature.append((key, inputs[key]))
|
|
|
|
return signature
|
|
|
|
|
|
|
|
# This function returns a list of all ancestors of the given node. The order of the list is
|
|
|
|
# deterministic based on which specific inputs the ancestor is connected by.
|
|
|
|
def get_ordered_ancestry(self, dynprompt, node_id):
|
|
|
|
ancestors = []
|
|
|
|
order_mapping = {}
|
|
|
|
self.get_ordered_ancestry_internal(dynprompt, node_id, ancestors, order_mapping)
|
|
|
|
return ancestors, order_mapping
|
|
|
|
|
|
|
|
def get_ordered_ancestry_internal(self, dynprompt, node_id, ancestors, order_mapping):
|
|
|
|
inputs = dynprompt.get_node(node_id)["inputs"]
|
|
|
|
input_keys = sorted(inputs.keys())
|
|
|
|
for key in input_keys:
|
|
|
|
if is_link(inputs[key]):
|
|
|
|
ancestor_id = inputs[key][0]
|
|
|
|
if ancestor_id not in order_mapping:
|
|
|
|
ancestors.append(ancestor_id)
|
|
|
|
order_mapping[ancestor_id] = len(ancestors) - 1
|
|
|
|
self.get_ordered_ancestry_internal(dynprompt, ancestor_id, ancestors, order_mapping)
|
|
|
|
|
|
|
|
class BasicCache:
|
|
|
|
def __init__(self, key_class):
|
|
|
|
self.key_class = key_class
|
|
|
|
self.initialized = False
|
|
|
|
self.dynprompt: DynamicPrompt
|
|
|
|
self.cache_key_set: CacheKeySet
|
|
|
|
self.cache = {}
|
|
|
|
self.subcaches = {}
|
|
|
|
|
|
|
|
def set_prompt(self, dynprompt, node_ids, is_changed_cache):
|
|
|
|
self.dynprompt = dynprompt
|
|
|
|
self.cache_key_set = self.key_class(dynprompt, node_ids, is_changed_cache)
|
|
|
|
self.is_changed_cache = is_changed_cache
|
|
|
|
self.initialized = True
|
|
|
|
|
|
|
|
def all_node_ids(self):
|
|
|
|
assert self.initialized
|
|
|
|
node_ids = self.cache_key_set.all_node_ids()
|
|
|
|
for subcache in self.subcaches.values():
|
|
|
|
node_ids = node_ids.union(subcache.all_node_ids())
|
|
|
|
return node_ids
|
|
|
|
|
|
|
|
def _clean_cache(self):
|
|
|
|
preserve_keys = set(self.cache_key_set.get_used_keys())
|
|
|
|
to_remove = []
|
|
|
|
for key in self.cache:
|
|
|
|
if key not in preserve_keys:
|
|
|
|
to_remove.append(key)
|
|
|
|
for key in to_remove:
|
|
|
|
del self.cache[key]
|
|
|
|
|
|
|
|
def _clean_subcaches(self):
|
|
|
|
preserve_subcaches = set(self.cache_key_set.get_used_subcache_keys())
|
|
|
|
|
|
|
|
to_remove = []
|
|
|
|
for key in self.subcaches:
|
|
|
|
if key not in preserve_subcaches:
|
|
|
|
to_remove.append(key)
|
|
|
|
for key in to_remove:
|
|
|
|
del self.subcaches[key]
|
|
|
|
|
|
|
|
def clean_unused(self):
|
|
|
|
assert self.initialized
|
|
|
|
self._clean_cache()
|
|
|
|
self._clean_subcaches()
|
|
|
|
|
|
|
|
def _set_immediate(self, node_id, value):
|
|
|
|
assert self.initialized
|
|
|
|
cache_key = self.cache_key_set.get_data_key(node_id)
|
|
|
|
self.cache[cache_key] = value
|
|
|
|
|
|
|
|
def _get_immediate(self, node_id):
|
|
|
|
if not self.initialized:
|
|
|
|
return None
|
|
|
|
cache_key = self.cache_key_set.get_data_key(node_id)
|
|
|
|
if cache_key in self.cache:
|
|
|
|
return self.cache[cache_key]
|
|
|
|
else:
|
|
|
|
return None
|
|
|
|
|
|
|
|
def _ensure_subcache(self, node_id, children_ids):
|
|
|
|
subcache_key = self.cache_key_set.get_subcache_key(node_id)
|
|
|
|
subcache = self.subcaches.get(subcache_key, None)
|
|
|
|
if subcache is None:
|
|
|
|
subcache = BasicCache(self.key_class)
|
|
|
|
self.subcaches[subcache_key] = subcache
|
|
|
|
subcache.set_prompt(self.dynprompt, children_ids, self.is_changed_cache)
|
|
|
|
return subcache
|
|
|
|
|
|
|
|
def _get_subcache(self, node_id):
|
|
|
|
assert self.initialized
|
|
|
|
subcache_key = self.cache_key_set.get_subcache_key(node_id)
|
|
|
|
if subcache_key in self.subcaches:
|
|
|
|
return self.subcaches[subcache_key]
|
|
|
|
else:
|
|
|
|
return None
|
|
|
|
|
|
|
|
def recursive_debug_dump(self):
|
|
|
|
result = []
|
|
|
|
for key in self.cache:
|
|
|
|
result.append({"key": key, "value": self.cache[key]})
|
|
|
|
for key in self.subcaches:
|
|
|
|
result.append({"subcache_key": key, "subcache": self.subcaches[key].recursive_debug_dump()})
|
|
|
|
return result
|
|
|
|
|
|
|
|
class HierarchicalCache(BasicCache):
|
|
|
|
def __init__(self, key_class):
|
|
|
|
super().__init__(key_class)
|
|
|
|
|
|
|
|
def _get_cache_for(self, node_id):
|
|
|
|
assert self.dynprompt is not None
|
|
|
|
parent_id = self.dynprompt.get_parent_node_id(node_id)
|
|
|
|
if parent_id is None:
|
|
|
|
return self
|
|
|
|
|
|
|
|
hierarchy = []
|
|
|
|
while parent_id is not None:
|
|
|
|
hierarchy.append(parent_id)
|
|
|
|
parent_id = self.dynprompt.get_parent_node_id(parent_id)
|
|
|
|
|
|
|
|
cache = self
|
|
|
|
for parent_id in reversed(hierarchy):
|
|
|
|
cache = cache._get_subcache(parent_id)
|
|
|
|
if cache is None:
|
|
|
|
return None
|
|
|
|
return cache
|
|
|
|
|
|
|
|
def get(self, node_id):
|
|
|
|
cache = self._get_cache_for(node_id)
|
|
|
|
if cache is None:
|
|
|
|
return None
|
|
|
|
return cache._get_immediate(node_id)
|
|
|
|
|
|
|
|
def set(self, node_id, value):
|
|
|
|
cache = self._get_cache_for(node_id)
|
|
|
|
assert cache is not None
|
|
|
|
cache._set_immediate(node_id, value)
|
|
|
|
|
|
|
|
def ensure_subcache_for(self, node_id, children_ids):
|
|
|
|
cache = self._get_cache_for(node_id)
|
|
|
|
assert cache is not None
|
|
|
|
return cache._ensure_subcache(node_id, children_ids)
|
|
|
|
|
|
|
|
class LRUCache(BasicCache):
|
|
|
|
def __init__(self, key_class, max_size=100):
|
|
|
|
super().__init__(key_class)
|
|
|
|
self.max_size = max_size
|
|
|
|
self.min_generation = 0
|
|
|
|
self.generation = 0
|
|
|
|
self.used_generation = {}
|
|
|
|
self.children = {}
|
|
|
|
|
|
|
|
def set_prompt(self, dynprompt, node_ids, is_changed_cache):
|
|
|
|
super().set_prompt(dynprompt, node_ids, is_changed_cache)
|
|
|
|
self.generation += 1
|
|
|
|
for node_id in node_ids:
|
|
|
|
self._mark_used(node_id)
|
|
|
|
|
|
|
|
def clean_unused(self):
|
|
|
|
while len(self.cache) > self.max_size and self.min_generation < self.generation:
|
|
|
|
self.min_generation += 1
|
|
|
|
to_remove = [key for key in self.cache if self.used_generation[key] < self.min_generation]
|
|
|
|
for key in to_remove:
|
|
|
|
del self.cache[key]
|
|
|
|
del self.used_generation[key]
|
|
|
|
if key in self.children:
|
|
|
|
del self.children[key]
|
|
|
|
self._clean_subcaches()
|
|
|
|
|
|
|
|
def get(self, node_id):
|
|
|
|
self._mark_used(node_id)
|
|
|
|
return self._get_immediate(node_id)
|
|
|
|
|
|
|
|
def _mark_used(self, node_id):
|
|
|
|
cache_key = self.cache_key_set.get_data_key(node_id)
|
|
|
|
if cache_key is not None:
|
|
|
|
self.used_generation[cache_key] = self.generation
|
|
|
|
|
|
|
|
def set(self, node_id, value):
|
|
|
|
self._mark_used(node_id)
|
|
|
|
return self._set_immediate(node_id, value)
|
|
|
|
|
|
|
|
def ensure_subcache_for(self, node_id, children_ids):
|
|
|
|
# Just uses subcaches for tracking 'live' nodes
|
|
|
|
super()._ensure_subcache(node_id, children_ids)
|
|
|
|
|
|
|
|
self.cache_key_set.add_keys(children_ids)
|
|
|
|
self._mark_used(node_id)
|
|
|
|
cache_key = self.cache_key_set.get_data_key(node_id)
|
|
|
|
self.children[cache_key] = []
|
|
|
|
for child_id in children_ids:
|
|
|
|
self._mark_used(child_id)
|
|
|
|
self.children[cache_key].append(self.cache_key_set.get_data_key(child_id))
|
|
|
|
return self
|
|
|
|
|