2022-12-01 05:33:20 +00:00
|
|
|
from .test_invoker import create_edge
|
|
|
|
from .test_nodes import ImageTestInvocation, ListPassThroughInvocation, PromptTestInvocation, PromptCollectionTestInvocation
|
2023-03-03 05:02:15 +00:00
|
|
|
from invokeai.app.invocations.baseinvocation import BaseInvocation, BaseInvocationOutput, InvocationContext
|
2023-04-06 04:06:05 +00:00
|
|
|
from invokeai.app.invocations.collections import RangeInvocation
|
|
|
|
from invokeai.app.invocations.math import AddInvocation, MultiplyInvocation
|
2023-03-03 05:02:15 +00:00
|
|
|
from invokeai.app.services.processor import DefaultInvocationProcessor
|
|
|
|
from invokeai.app.services.sqlite import SqliteItemStorage, sqlite_memory
|
|
|
|
from invokeai.app.services.invocation_queue import MemoryInvocationQueue
|
|
|
|
from invokeai.app.services.invocation_services import InvocationServices
|
|
|
|
from invokeai.app.services.graph import Graph, GraphInvocation, InvalidEdgeError, NodeAlreadyInGraphError, NodeNotFoundError, are_connections_compatible, EdgeConnection, CollectInvocation, IterateInvocation, GraphExecutionState
|
2022-12-01 05:33:20 +00:00
|
|
|
import pytest
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.fixture
|
|
|
|
def simple_graph():
|
|
|
|
g = Graph()
|
|
|
|
g.add_node(PromptTestInvocation(id = "1", prompt = "Banana sushi"))
|
|
|
|
g.add_node(ImageTestInvocation(id = "2"))
|
|
|
|
g.add_edge(create_edge("1", "prompt", "2", "prompt"))
|
|
|
|
return g
|
|
|
|
|
|
|
|
@pytest.fixture
|
|
|
|
def mock_services():
|
|
|
|
# NOTE: none of these are actually called by the test invocations
|
2023-02-25 04:11:28 +00:00
|
|
|
return InvocationServices(
|
2023-04-06 04:06:05 +00:00
|
|
|
model_manager = None, # type: ignore
|
|
|
|
events = None, # type: ignore
|
|
|
|
images = None, # type: ignore
|
|
|
|
latents = None, # type: ignore
|
2023-02-25 04:11:28 +00:00
|
|
|
queue = MemoryInvocationQueue(),
|
|
|
|
graph_execution_manager = SqliteItemStorage[GraphExecutionState](filename = sqlite_memory, table_name = 'graph_executions'),
|
2023-03-11 22:00:00 +00:00
|
|
|
processor = DefaultInvocationProcessor(),
|
2023-04-06 04:06:05 +00:00
|
|
|
restoration = None, # type: ignore
|
2023-02-25 04:11:28 +00:00
|
|
|
)
|
2022-12-01 05:33:20 +00:00
|
|
|
|
|
|
|
def invoke_next(g: GraphExecutionState, services: InvocationServices) -> tuple[BaseInvocation, BaseInvocationOutput]:
|
|
|
|
n = g.next()
|
|
|
|
if n is None:
|
|
|
|
return (None, None)
|
|
|
|
|
|
|
|
print(f'invoking {n.id}: {type(n)}')
|
|
|
|
o = n.invoke(InvocationContext(services, "1"))
|
|
|
|
g.complete(n.id, o)
|
|
|
|
|
|
|
|
return (n, o)
|
|
|
|
|
|
|
|
def test_graph_state_executes_in_order(simple_graph, mock_services):
|
|
|
|
g = GraphExecutionState(graph = simple_graph)
|
|
|
|
|
|
|
|
n1 = invoke_next(g, mock_services)
|
|
|
|
n2 = invoke_next(g, mock_services)
|
|
|
|
n3 = g.next()
|
|
|
|
|
|
|
|
assert g.prepared_source_mapping[n1[0].id] == "1"
|
|
|
|
assert g.prepared_source_mapping[n2[0].id] == "2"
|
|
|
|
assert n3 is None
|
|
|
|
assert g.results[n1[0].id].prompt == n1[0].prompt
|
|
|
|
assert n2[0].prompt == n1[0].prompt
|
|
|
|
|
|
|
|
def test_graph_is_complete(simple_graph, mock_services):
|
|
|
|
g = GraphExecutionState(graph = simple_graph)
|
|
|
|
n1 = invoke_next(g, mock_services)
|
|
|
|
n2 = invoke_next(g, mock_services)
|
|
|
|
n3 = g.next()
|
|
|
|
|
|
|
|
assert g.is_complete()
|
|
|
|
|
|
|
|
def test_graph_is_not_complete(simple_graph, mock_services):
|
|
|
|
g = GraphExecutionState(graph = simple_graph)
|
|
|
|
n1 = invoke_next(g, mock_services)
|
|
|
|
n2 = g.next()
|
|
|
|
|
|
|
|
assert not g.is_complete()
|
|
|
|
|
|
|
|
# TODO: test completion with iterators/subgraphs
|
|
|
|
|
|
|
|
def test_graph_state_expands_iterator(mock_services):
|
|
|
|
graph = Graph()
|
2023-04-06 04:06:05 +00:00
|
|
|
graph.add_node(RangeInvocation(id = "0", start = 0, stop = 3, step = 1))
|
|
|
|
graph.add_node(IterateInvocation(id = "1"))
|
|
|
|
graph.add_node(MultiplyInvocation(id = "2", b = 10))
|
|
|
|
graph.add_node(AddInvocation(id = "3", b = 1))
|
|
|
|
graph.add_edge(create_edge("0", "collection", "1", "collection"))
|
|
|
|
graph.add_edge(create_edge("1", "item", "2", "a"))
|
|
|
|
graph.add_edge(create_edge("2", "a", "3", "a"))
|
2022-12-01 05:33:20 +00:00
|
|
|
|
|
|
|
g = GraphExecutionState(graph = graph)
|
2023-04-06 04:06:05 +00:00
|
|
|
while not g.is_complete():
|
|
|
|
invoke_next(g, mock_services)
|
|
|
|
|
|
|
|
prepared_add_nodes = g.source_prepared_mapping['3']
|
|
|
|
results = set([g.results[n].a for n in prepared_add_nodes])
|
|
|
|
expected = set([1, 11, 21])
|
|
|
|
assert results == expected
|
2022-12-01 05:33:20 +00:00
|
|
|
|
|
|
|
|
|
|
|
def test_graph_state_collects(mock_services):
|
|
|
|
graph = Graph()
|
|
|
|
test_prompts = ["Banana sushi", "Cat sushi"]
|
|
|
|
graph.add_node(PromptCollectionTestInvocation(id = "1", collection = list(test_prompts)))
|
|
|
|
graph.add_node(IterateInvocation(id = "2"))
|
|
|
|
graph.add_node(PromptTestInvocation(id = "3"))
|
|
|
|
graph.add_node(CollectInvocation(id = "4"))
|
|
|
|
graph.add_edge(create_edge("1", "collection", "2", "collection"))
|
|
|
|
graph.add_edge(create_edge("2", "item", "3", "prompt"))
|
|
|
|
graph.add_edge(create_edge("3", "prompt", "4", "item"))
|
|
|
|
|
|
|
|
g = GraphExecutionState(graph = graph)
|
|
|
|
n1 = invoke_next(g, mock_services)
|
|
|
|
n2 = invoke_next(g, mock_services)
|
|
|
|
n3 = invoke_next(g, mock_services)
|
|
|
|
n4 = invoke_next(g, mock_services)
|
|
|
|
n5 = invoke_next(g, mock_services)
|
|
|
|
n6 = invoke_next(g, mock_services)
|
|
|
|
|
|
|
|
assert isinstance(n6[0], CollectInvocation)
|
|
|
|
|
|
|
|
assert sorted(g.results[n6[0].id].collection) == sorted(test_prompts)
|