Group nodes (#1776)
* setup ui unit tests * Refactoring, adding connections * Few tweaks * Fix type * Add general test * Refactored and extended test * move to describe * for groups * wip group nodes * Relink nodes Fixed widget values Convert to nodes * Reconnect on convert back * add via node menu + canvas refactor * Add ws event handling * fix using wrong node on widget serialize * allow reroute pipe fix control_after_generate configure * allow multiple images * Add test for converted widgets on missing nodes + fix crash * tidy * mores tests + refactor * throw earlier to get less confusing error * support outputs * more test * add ci action * use lts node * Fix? * Prevent connecting non matching combos * update * accidently removed npm i * Disable logging extension * fix naming allow control_after_generate custom name allow convert from reroutes * group node tests * Add executing info, custom node icon Tidy * internal reroute just works * Fix crash on virtual nodes e.g. note * Save group nodes to templates * Fix template nodes not being stored * Fix aborting convert * tidy * Fix reconnecting output links on convert to group * Fix links on convert to nodes * Handle missing internal nodes * Trigger callback on text change * Apply value on connect * Fix converted widgets not reconnecting * Group node updates - persist internal ids in current session - copy widget values when converting to nodes - fix issue serializing converted inputs * Resolve issue with sanitized node name * Fix internal id * allow outputs to be used internally and externally * order widgets on group node various fixes * fix imageupload widget requiring a specific name * groupnode imageupload test give widget unique name * Fix issue with external node links * Add VAE model * Fix internal node id check * fix potential crash * wip widget input support * more wip group widget inputs * Group node refactor Support for primitives/converted widgets * Fix convert to nodes with internal reroutes * fix applying primitive * Fix control widget values * fix testmain
parent
d19de2753e
commit
7f469203b7
@ -0,0 +1,9 @@
|
||||
{
|
||||
"path-intellisense.mappings": {
|
||||
"../": "${workspaceFolder}/web/extensions/core"
|
||||
},
|
||||
"[python]": {
|
||||
"editor.defaultFormatter": "ms-python.autopep8"
|
||||
},
|
||||
"python.formatting.provider": "none"
|
||||
}
|
@ -0,0 +1,818 @@
|
||||
// @ts-check
|
||||
/// <reference path="../node_modules/@types/jest/index.d.ts" />
|
||||
|
||||
const { start, createDefaultWorkflow } = require("../utils");
|
||||
const lg = require("../utils/litegraph");
|
||||
|
||||
describe("group node", () => {
|
||||
beforeEach(() => {
|
||||
lg.setup(global);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
lg.teardown(global);
|
||||
});
|
||||
|
||||
/**
|
||||
*
|
||||
* @param {*} app
|
||||
* @param {*} graph
|
||||
* @param {*} name
|
||||
* @param {*} nodes
|
||||
* @returns { Promise<InstanceType<import("../utils/ezgraph")["EzNode"]>> }
|
||||
*/
|
||||
async function convertToGroup(app, graph, name, nodes) {
|
||||
// Select the nodes we are converting
|
||||
for (const n of nodes) {
|
||||
n.select(true);
|
||||
}
|
||||
|
||||
expect(Object.keys(app.canvas.selected_nodes).sort((a, b) => +a - +b)).toEqual(
|
||||
nodes.map((n) => n.id + "").sort((a, b) => +a - +b)
|
||||
);
|
||||
|
||||
global.prompt = jest.fn().mockImplementation(() => name);
|
||||
const groupNode = await nodes[0].menu["Convert to Group Node"].call(false);
|
||||
|
||||
// Check group name was requested
|
||||
expect(window.prompt).toHaveBeenCalled();
|
||||
|
||||
// Ensure old nodes are removed
|
||||
for (const n of nodes) {
|
||||
expect(n.isRemoved).toBeTruthy();
|
||||
}
|
||||
|
||||
expect(groupNode.type).toEqual("workflow/" + name);
|
||||
|
||||
return graph.find(groupNode);
|
||||
}
|
||||
|
||||
/**
|
||||
* @param { Record<string, string | number> | number[] } idMap
|
||||
* @param { Record<string, Record<string, unknown>> } valueMap
|
||||
*/
|
||||
function getOutput(idMap = {}, valueMap = {}) {
|
||||
if (idMap instanceof Array) {
|
||||
idMap = idMap.reduce((p, n) => {
|
||||
p[n] = n + "";
|
||||
return p;
|
||||
}, {});
|
||||
}
|
||||
const expected = {
|
||||
1: { inputs: { ckpt_name: "model1.safetensors", ...valueMap?.[1] }, class_type: "CheckpointLoaderSimple" },
|
||||
2: { inputs: { text: "positive", clip: ["1", 1], ...valueMap?.[2] }, class_type: "CLIPTextEncode" },
|
||||
3: { inputs: { text: "negative", clip: ["1", 1], ...valueMap?.[3] }, class_type: "CLIPTextEncode" },
|
||||
4: { inputs: { width: 512, height: 512, batch_size: 1, ...valueMap?.[4] }, class_type: "EmptyLatentImage" },
|
||||
5: {
|
||||
inputs: {
|
||||
seed: 0,
|
||||
steps: 20,
|
||||
cfg: 8,
|
||||
sampler_name: "euler",
|
||||
scheduler: "normal",
|
||||
denoise: 1,
|
||||
model: ["1", 0],
|
||||
positive: ["2", 0],
|
||||
negative: ["3", 0],
|
||||
latent_image: ["4", 0],
|
||||
...valueMap?.[5],
|
||||
},
|
||||
class_type: "KSampler",
|
||||
},
|
||||
6: { inputs: { samples: ["5", 0], vae: ["1", 2], ...valueMap?.[6] }, class_type: "VAEDecode" },
|
||||
7: { inputs: { filename_prefix: "ComfyUI", images: ["6", 0], ...valueMap?.[7] }, class_type: "SaveImage" },
|
||||
};
|
||||
|
||||
// Map old IDs to new at the top level
|
||||
const mapped = {};
|
||||
for (const oldId in idMap) {
|
||||
mapped[idMap[oldId]] = expected[oldId];
|
||||
delete expected[oldId];
|
||||
}
|
||||
Object.assign(mapped, expected);
|
||||
|
||||
// Map old IDs to new inside links
|
||||
for (const k in mapped) {
|
||||
for (const input in mapped[k].inputs) {
|
||||
const v = mapped[k].inputs[input];
|
||||
if (v instanceof Array) {
|
||||
if (v[0] in idMap) {
|
||||
v[0] = idMap[v[0]] + "";
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return mapped;
|
||||
}
|
||||
|
||||
test("can be created from selected nodes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg, nodes.empty]);
|
||||
|
||||
// Ensure links are now to the group node
|
||||
expect(group.inputs).toHaveLength(2);
|
||||
expect(group.outputs).toHaveLength(3);
|
||||
|
||||
expect(group.inputs.map((i) => i.input.name)).toEqual(["clip", "CLIPTextEncode clip"]);
|
||||
expect(group.outputs.map((i) => i.output.name)).toEqual(["LATENT", "CONDITIONING", "CLIPTextEncode CONDITIONING"]);
|
||||
|
||||
// ckpt clip to both clip inputs on the group
|
||||
expect(nodes.ckpt.outputs.CLIP.connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[group.id, 0],
|
||||
[group.id, 1],
|
||||
]);
|
||||
|
||||
// group conditioning to sampler
|
||||
expect(group.outputs["CONDITIONING"].connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[nodes.sampler.id, 1],
|
||||
]);
|
||||
// group conditioning 2 to sampler
|
||||
expect(
|
||||
group.outputs["CLIPTextEncode CONDITIONING"].connections.map((t) => [t.targetNode.id, t.targetInput.index])
|
||||
).toEqual([[nodes.sampler.id, 2]]);
|
||||
// group latent to sampler
|
||||
expect(group.outputs["LATENT"].connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[nodes.sampler.id, 3],
|
||||
]);
|
||||
});
|
||||
|
||||
test("maintains all output links on conversion", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const save2 = ez.SaveImage(...nodes.decode.outputs);
|
||||
const save3 = ez.SaveImage(...nodes.decode.outputs);
|
||||
// Ensure an output with multiple links maintains them on convert to group
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.sampler, nodes.decode]);
|
||||
expect(group.outputs[0].connections.length).toBe(3);
|
||||
expect(group.outputs[0].connections[0].targetNode.id).toBe(nodes.save.id);
|
||||
expect(group.outputs[0].connections[1].targetNode.id).toBe(save2.id);
|
||||
expect(group.outputs[0].connections[2].targetNode.id).toBe(save3.id);
|
||||
|
||||
// and they're still linked when converting back to nodes
|
||||
const newNodes = group.menu["Convert to nodes"].call();
|
||||
const decode = graph.find(newNodes.find((n) => n.type === "VAEDecode"));
|
||||
expect(decode.outputs[0].connections.length).toBe(3);
|
||||
expect(decode.outputs[0].connections[0].targetNode.id).toBe(nodes.save.id);
|
||||
expect(decode.outputs[0].connections[1].targetNode.id).toBe(save2.id);
|
||||
expect(decode.outputs[0].connections[2].targetNode.id).toBe(save3.id);
|
||||
});
|
||||
test("can be be converted back to nodes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const toConvert = [nodes.pos, nodes.neg, nodes.empty, nodes.sampler];
|
||||
const group = await convertToGroup(app, graph, "test", toConvert);
|
||||
|
||||
// Edit some values to ensure they are set back onto the converted nodes
|
||||
expect(group.widgets["text"].value).toBe("positive");
|
||||
group.widgets["text"].value = "pos";
|
||||
expect(group.widgets["CLIPTextEncode text"].value).toBe("negative");
|
||||
group.widgets["CLIPTextEncode text"].value = "neg";
|
||||
expect(group.widgets["width"].value).toBe(512);
|
||||
group.widgets["width"].value = 1024;
|
||||
expect(group.widgets["sampler_name"].value).toBe("euler");
|
||||
group.widgets["sampler_name"].value = "ddim";
|
||||
expect(group.widgets["control_after_generate"].value).toBe("randomize");
|
||||
group.widgets["control_after_generate"].value = "fixed";
|
||||
|
||||
/** @type { Array<any> } */
|
||||
group.menu["Convert to nodes"].call();
|
||||
|
||||
// ensure widget values are set
|
||||
const pos = graph.find(nodes.pos.id);
|
||||
expect(pos.node.type).toBe("CLIPTextEncode");
|
||||
expect(pos.widgets["text"].value).toBe("pos");
|
||||
const neg = graph.find(nodes.neg.id);
|
||||
expect(neg.node.type).toBe("CLIPTextEncode");
|
||||
expect(neg.widgets["text"].value).toBe("neg");
|
||||
const empty = graph.find(nodes.empty.id);
|
||||
expect(empty.node.type).toBe("EmptyLatentImage");
|
||||
expect(empty.widgets["width"].value).toBe(1024);
|
||||
const sampler = graph.find(nodes.sampler.id);
|
||||
expect(sampler.node.type).toBe("KSampler");
|
||||
expect(sampler.widgets["sampler_name"].value).toBe("ddim");
|
||||
expect(sampler.widgets["control_after_generate"].value).toBe("fixed");
|
||||
|
||||
// validate links
|
||||
expect(nodes.ckpt.outputs.CLIP.connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[pos.id, 0],
|
||||
[neg.id, 0],
|
||||
]);
|
||||
|
||||
expect(pos.outputs["CONDITIONING"].connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[nodes.sampler.id, 1],
|
||||
]);
|
||||
|
||||
expect(neg.outputs["CONDITIONING"].connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[nodes.sampler.id, 2],
|
||||
]);
|
||||
|
||||
expect(empty.outputs["LATENT"].connections.map((t) => [t.targetNode.id, t.targetInput.index])).toEqual([
|
||||
[nodes.sampler.id, 3],
|
||||
]);
|
||||
});
|
||||
test("it can embed reroutes as inputs", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
|
||||
// Add and connect a reroute to the clip text encodes
|
||||
const reroute = ez.Reroute();
|
||||
nodes.ckpt.outputs.CLIP.connectTo(reroute.inputs[0]);
|
||||
reroute.outputs[0].connectTo(nodes.pos.inputs[0]);
|
||||
reroute.outputs[0].connectTo(nodes.neg.inputs[0]);
|
||||
|
||||
// Convert to group and ensure we only have 1 input of the correct type
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg, nodes.empty, reroute]);
|
||||
expect(group.inputs).toHaveLength(1);
|
||||
expect(group.inputs[0].input.type).toEqual("CLIP");
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual(getOutput());
|
||||
});
|
||||
test("it can embed reroutes as outputs", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
|
||||
// Add a reroute with no output so we output IMAGE even though its used internally
|
||||
const reroute = ez.Reroute();
|
||||
nodes.decode.outputs.IMAGE.connectTo(reroute.inputs[0]);
|
||||
|
||||
// Convert to group and ensure there is an IMAGE output
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.decode, nodes.save, reroute]);
|
||||
expect(group.outputs).toHaveLength(1);
|
||||
expect(group.outputs[0].output.type).toEqual("IMAGE");
|
||||
expect((await graph.toPrompt()).output).toEqual(getOutput([nodes.decode.id, nodes.save.id]));
|
||||
});
|
||||
test("it can embed reroutes as pipes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
|
||||
// Use reroutes as a pipe
|
||||
const rerouteModel = ez.Reroute();
|
||||
const rerouteClip = ez.Reroute();
|
||||
const rerouteVae = ez.Reroute();
|
||||
nodes.ckpt.outputs.MODEL.connectTo(rerouteModel.inputs[0]);
|
||||
nodes.ckpt.outputs.CLIP.connectTo(rerouteClip.inputs[0]);
|
||||
nodes.ckpt.outputs.VAE.connectTo(rerouteVae.inputs[0]);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [rerouteModel, rerouteClip, rerouteVae]);
|
||||
|
||||
expect(group.outputs).toHaveLength(3);
|
||||
expect(group.outputs.map((o) => o.output.type)).toEqual(["MODEL", "CLIP", "VAE"]);
|
||||
|
||||
expect(group.outputs).toHaveLength(3);
|
||||
expect(group.outputs.map((o) => o.output.type)).toEqual(["MODEL", "CLIP", "VAE"]);
|
||||
|
||||
group.outputs[0].connectTo(nodes.sampler.inputs.model);
|
||||
group.outputs[1].connectTo(nodes.pos.inputs.clip);
|
||||
group.outputs[1].connectTo(nodes.neg.inputs.clip);
|
||||
});
|
||||
test("can handle reroutes used internally", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
|
||||
let reroutes = [];
|
||||
let prevNode = nodes.ckpt;
|
||||
for(let i = 0; i < 5; i++) {
|
||||
const reroute = ez.Reroute();
|
||||
prevNode.outputs[0].connectTo(reroute.inputs[0]);
|
||||
prevNode = reroute;
|
||||
reroutes.push(reroute);
|
||||
}
|
||||
prevNode.outputs[0].connectTo(nodes.sampler.inputs.model);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [...reroutes, ...Object.values(nodes)]);
|
||||
expect((await graph.toPrompt()).output).toEqual(getOutput());
|
||||
|
||||
group.menu["Convert to nodes"].call();
|
||||
expect((await graph.toPrompt()).output).toEqual(getOutput());
|
||||
});
|
||||
test("creates with widget values from inner nodes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
|
||||
nodes.ckpt.widgets.ckpt_name.value = "model2.ckpt";
|
||||
nodes.pos.widgets.text.value = "hello";
|
||||
nodes.neg.widgets.text.value = "world";
|
||||
nodes.empty.widgets.width.value = 256;
|
||||
nodes.empty.widgets.height.value = 1024;
|
||||
nodes.sampler.widgets.seed.value = 1;
|
||||
nodes.sampler.widgets.control_after_generate.value = "increment";
|
||||
nodes.sampler.widgets.steps.value = 8;
|
||||
nodes.sampler.widgets.cfg.value = 4.5;
|
||||
nodes.sampler.widgets.sampler_name.value = "uni_pc";
|
||||
nodes.sampler.widgets.scheduler.value = "karras";
|
||||
nodes.sampler.widgets.denoise.value = 0.9;
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [
|
||||
nodes.ckpt,
|
||||
nodes.pos,
|
||||
nodes.neg,
|
||||
nodes.empty,
|
||||
nodes.sampler,
|
||||
]);
|
||||
|
||||
expect(group.widgets["ckpt_name"].value).toEqual("model2.ckpt");
|
||||
expect(group.widgets["text"].value).toEqual("hello");
|
||||
expect(group.widgets["CLIPTextEncode text"].value).toEqual("world");
|
||||
expect(group.widgets["width"].value).toEqual(256);
|
||||
expect(group.widgets["height"].value).toEqual(1024);
|
||||
expect(group.widgets["seed"].value).toEqual(1);
|
||||
expect(group.widgets["control_after_generate"].value).toEqual("increment");
|
||||
expect(group.widgets["steps"].value).toEqual(8);
|
||||
expect(group.widgets["cfg"].value).toEqual(4.5);
|
||||
expect(group.widgets["sampler_name"].value).toEqual("uni_pc");
|
||||
expect(group.widgets["scheduler"].value).toEqual("karras");
|
||||
expect(group.widgets["denoise"].value).toEqual(0.9);
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual(
|
||||
getOutput([nodes.ckpt.id, nodes.pos.id, nodes.neg.id, nodes.empty.id, nodes.sampler.id], {
|
||||
[nodes.ckpt.id]: { ckpt_name: "model2.ckpt" },
|
||||
[nodes.pos.id]: { text: "hello" },
|
||||
[nodes.neg.id]: { text: "world" },
|
||||
[nodes.empty.id]: { width: 256, height: 1024 },
|
||||
[nodes.sampler.id]: {
|
||||
seed: 1,
|
||||
steps: 8,
|
||||
cfg: 4.5,
|
||||
sampler_name: "uni_pc",
|
||||
scheduler: "karras",
|
||||
denoise: 0.9,
|
||||
},
|
||||
})
|
||||
);
|
||||
});
|
||||
test("group inputs can be reroutes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg]);
|
||||
|
||||
const reroute = ez.Reroute();
|
||||
nodes.ckpt.outputs.CLIP.connectTo(reroute.inputs[0]);
|
||||
|
||||
reroute.outputs[0].connectTo(group.inputs[0]);
|
||||
reroute.outputs[0].connectTo(group.inputs[1]);
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual(getOutput([nodes.pos.id, nodes.neg.id]));
|
||||
});
|
||||
test("group outputs can be reroutes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg]);
|
||||
|
||||
const reroute1 = ez.Reroute();
|
||||
const reroute2 = ez.Reroute();
|
||||
group.outputs[0].connectTo(reroute1.inputs[0]);
|
||||
group.outputs[1].connectTo(reroute2.inputs[0]);
|
||||
|
||||
reroute1.outputs[0].connectTo(nodes.sampler.inputs.positive);
|
||||
reroute2.outputs[0].connectTo(nodes.sampler.inputs.negative);
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual(getOutput([nodes.pos.id, nodes.neg.id]));
|
||||
});
|
||||
test("groups can connect to each other", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const group1 = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg]);
|
||||
const group2 = await convertToGroup(app, graph, "test2", [nodes.empty, nodes.sampler]);
|
||||
|
||||
group1.outputs[0].connectTo(group2.inputs["positive"]);
|
||||
group1.outputs[1].connectTo(group2.inputs["negative"]);
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual(
|
||||
getOutput([nodes.pos.id, nodes.neg.id, nodes.empty.id, nodes.sampler.id])
|
||||
);
|
||||
});
|
||||
test("displays generated image on group node", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
let group = await convertToGroup(app, graph, "test", [
|
||||
nodes.pos,
|
||||
nodes.neg,
|
||||
nodes.empty,
|
||||
nodes.sampler,
|
||||
nodes.decode,
|
||||
nodes.save,
|
||||
]);
|
||||
|
||||
const { api } = require("../../web/scripts/api");
|
||||
|
||||
api.dispatchEvent(new CustomEvent("execution_start", {}));
|
||||
api.dispatchEvent(new CustomEvent("executing", { detail: `${nodes.save.id}` }));
|
||||
// Event should be forwarded to group node id
|
||||
expect(+app.runningNodeId).toEqual(group.id);
|
||||
expect(group.node["imgs"]).toBeFalsy();
|
||||
api.dispatchEvent(
|
||||
new CustomEvent("executed", {
|
||||
detail: {
|
||||
node: `${nodes.save.id}`,
|
||||
output: {
|
||||
images: [
|
||||
{
|
||||
filename: "test.png",
|
||||
type: "output",
|
||||
},
|
||||
],
|
||||
},
|
||||
},
|
||||
})
|
||||
);
|
||||
|
||||
// Trigger paint
|
||||
group.node.onDrawBackground?.(app.canvas.ctx, app.canvas.canvas);
|
||||
|
||||
expect(group.node["images"]).toEqual([
|
||||
{
|
||||
filename: "test.png",
|
||||
type: "output",
|
||||
},
|
||||
]);
|
||||
|
||||
// Reload
|
||||
const workflow = JSON.stringify((await graph.toPrompt()).workflow);
|
||||
await app.loadGraphData(JSON.parse(workflow));
|
||||
group = graph.find(group);
|
||||
|
||||
// Trigger inner nodes to get created
|
||||
group.node["getInnerNodes"]();
|
||||
|
||||
// Check it works for internal node ids
|
||||
api.dispatchEvent(new CustomEvent("execution_start", {}));
|
||||
api.dispatchEvent(new CustomEvent("executing", { detail: `${group.id}:5` }));
|
||||
// Event should be forwarded to group node id
|
||||
expect(+app.runningNodeId).toEqual(group.id);
|
||||
expect(group.node["imgs"]).toBeFalsy();
|
||||
api.dispatchEvent(
|
||||
new CustomEvent("executed", {
|
||||
detail: {
|
||||
node: `${group.id}:5`,
|
||||
output: {
|
||||
images: [
|
||||
{
|
||||
filename: "test2.png",
|
||||
type: "output",
|
||||
},
|
||||
],
|
||||
},
|
||||
},
|
||||
})
|
||||
);
|
||||
|
||||
// Trigger paint
|
||||
group.node.onDrawBackground?.(app.canvas.ctx, app.canvas.canvas);
|
||||
|
||||
expect(group.node["images"]).toEqual([
|
||||
{
|
||||
filename: "test2.png",
|
||||
type: "output",
|
||||
},
|
||||
]);
|
||||
});
|
||||
test("allows widgets to be converted to inputs", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
const group = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg]);
|
||||
group.widgets[0].convertToInput();
|
||||
|
||||
const primitive = ez.PrimitiveNode();
|
||||
primitive.outputs[0].connectTo(group.inputs["text"]);
|
||||
primitive.widgets[0].value = "hello";
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual(
|
||||
getOutput([nodes.pos.id, nodes.neg.id], {
|
||||
[nodes.pos.id]: { text: "hello" },
|
||||
})
|
||||
);
|
||||
});
|
||||
test("can be copied", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
|
||||
const group1 = await convertToGroup(app, graph, "test", [
|
||||
nodes.pos,
|
||||
nodes.neg,
|
||||
nodes.empty,
|
||||
nodes.sampler,
|
||||
nodes.decode,
|
||||
nodes.save,
|
||||
]);
|
||||
|
||||
group1.widgets["text"].value = "hello";
|
||||
group1.widgets["width"].value = 256;
|
||||
group1.widgets["seed"].value = 1;
|
||||
|
||||
// Clone the node
|
||||
group1.menu.Clone.call();
|
||||
expect(app.graph._nodes).toHaveLength(3);
|
||||
const group2 = graph.find(app.graph._nodes[2]);
|
||||
expect(group2.node.type).toEqual("workflow/test");
|
||||
expect(group2.id).not.toEqual(group1.id);
|
||||
|
||||
// Reconnect ckpt
|
||||
nodes.ckpt.outputs.MODEL.connectTo(group2.inputs["model"]);
|
||||
nodes.ckpt.outputs.CLIP.connectTo(group2.inputs["clip"]);
|
||||
nodes.ckpt.outputs.CLIP.connectTo(group2.inputs["CLIPTextEncode clip"]);
|
||||
nodes.ckpt.outputs.VAE.connectTo(group2.inputs["vae"]);
|
||||
|
||||
group2.widgets["text"].value = "world";
|
||||
group2.widgets["width"].value = 1024;
|
||||
group2.widgets["seed"].value = 100;
|
||||
|
||||
let i = 0;
|
||||
expect((await graph.toPrompt()).output).toEqual({
|
||||
...getOutput([nodes.empty.id, nodes.pos.id, nodes.neg.id, nodes.sampler.id, nodes.decode.id, nodes.save.id], {
|
||||
[nodes.empty.id]: { width: 256 },
|
||||
[nodes.pos.id]: { text: "hello" },
|
||||
[nodes.sampler.id]: { seed: 1 },
|
||||
}),
|
||||
...getOutput(
|
||||
{
|
||||
[nodes.empty.id]: `${group2.id}:${i++}`,
|
||||
[nodes.pos.id]: `${group2.id}:${i++}`,
|
||||
[nodes.neg.id]: `${group2.id}:${i++}`,
|
||||
[nodes.sampler.id]: `${group2.id}:${i++}`,
|
||||
[nodes.decode.id]: `${group2.id}:${i++}`,
|
||||
[nodes.save.id]: `${group2.id}:${i++}`,
|
||||
},
|
||||
{
|
||||
[nodes.empty.id]: { width: 1024 },
|
||||
[nodes.pos.id]: { text: "world" },
|
||||
[nodes.sampler.id]: { seed: 100 },
|
||||
}
|
||||
),
|
||||
});
|
||||
|
||||
graph.arrange();
|
||||
});
|
||||
test("is embedded in workflow", async () => {
|
||||
let { ez, graph, app } = await start();
|
||||
const nodes = createDefaultWorkflow(ez, graph);
|
||||
let group = await convertToGroup(app, graph, "test", [nodes.pos, nodes.neg]);
|
||||
const workflow = JSON.stringify((await graph.toPrompt()).workflow);
|
||||
|
||||
// Clear the environment
|
||||
({ ez, graph, app } = await start({
|
||||
resetEnv: true,
|
||||
}));
|
||||
// Ensure the node isnt registered
|
||||
expect(() => ez["workflow/test"]).toThrow();
|
||||
|
||||
// Reload the workflow
|
||||
await app.loadGraphData(JSON.parse(workflow));
|
||||
|
||||
// Ensure the node is found
|
||||
group = graph.find(group);
|
||||
|
||||
// Generate prompt and ensure it is as expected
|
||||
expect((await graph.toPrompt()).output).toEqual(
|
||||
getOutput({
|
||||
[nodes.pos.id]: `${group.id}:0`,
|
||||
[nodes.neg.id]: `${group.id}:1`,
|
||||
})
|
||||
);
|
||||
});
|
||||
test("shows missing node error on missing internal node when loading graph data", async () => {
|
||||
const { graph } = await start();
|
||||
|
||||
const dialogShow = jest.spyOn(graph.app.ui.dialog, "show");
|
||||
await graph.app.loadGraphData({
|
||||
last_node_id: 3,
|
||||
last_link_id: 1,
|
||||
nodes: [
|
||||
{
|
||||
id: 3,
|
||||
type: "workflow/testerror",
|
||||
},
|
||||
],
|
||||
links: [],
|
||||
groups: [],
|
||||
config: {},
|
||||
extra: {
|
||||
groupNodes: {
|
||||
testerror: {
|
||||
nodes: [
|
||||
{
|
||||
type: "NotKSampler",
|
||||
},
|
||||
{
|
||||
type: "NotVAEDecode",
|
||||
},
|
||||
],
|
||||
},
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
expect(dialogShow).toBeCalledTimes(1);
|
||||
const call = dialogShow.mock.calls[0][0].innerHTML;
|
||||
expect(call).toContain("the following node types were not found");
|
||||
expect(call).toContain("NotKSampler");
|
||||
expect(call).toContain("NotVAEDecode");
|
||||
expect(call).toContain("workflow/testerror");
|
||||
});
|
||||
test("maintains widget inputs on conversion back to nodes", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
let pos = ez.CLIPTextEncode({ text: "positive" });
|
||||
pos.node.title = "Positive";
|
||||
let neg = ez.CLIPTextEncode({ text: "negative" });
|
||||
neg.node.title = "Negative";
|
||||
pos.widgets.text.convertToInput();
|
||||
neg.widgets.text.convertToInput();
|
||||
|
||||
let primitive = ez.PrimitiveNode();
|
||||
primitive.outputs[0].connectTo(pos.inputs.text);
|
||||
primitive.outputs[0].connectTo(neg.inputs.text);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [pos, neg, primitive]);
|
||||
// This will use a primitive widget named 'value'
|
||||
expect(group.widgets.length).toBe(1);
|
||||
expect(group.widgets["value"].value).toBe("positive");
|
||||
|
||||
const newNodes = group.menu["Convert to nodes"].call();
|
||||
pos = graph.find(newNodes.find((n) => n.title === "Positive"));
|
||||
neg = graph.find(newNodes.find((n) => n.title === "Negative"));
|
||||
primitive = graph.find(newNodes.find((n) => n.type === "PrimitiveNode"));
|
||||
|
||||
expect(pos.inputs).toHaveLength(2);
|
||||
expect(neg.inputs).toHaveLength(2);
|
||||
expect(primitive.outputs[0].connections).toHaveLength(2);
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual({
|
||||
1: { inputs: { text: "positive" }, class_type: "CLIPTextEncode" },
|
||||
2: { inputs: { text: "positive" }, class_type: "CLIPTextEncode" },
|
||||
});
|
||||
});
|
||||
test("adds widgets in node execution order", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const scale = ez.LatentUpscale();
|
||||
const save = ez.SaveImage();
|
||||
const empty = ez.EmptyLatentImage();
|
||||
const decode = ez.VAEDecode();
|
||||
|
||||
scale.outputs.LATENT.connectTo(decode.inputs.samples);
|
||||
decode.outputs.IMAGE.connectTo(save.inputs.images);
|
||||
empty.outputs.LATENT.connectTo(scale.inputs.samples);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [scale, save, empty, decode]);
|
||||
const widgets = group.widgets.map((w) => w.widget.name);
|
||||
expect(widgets).toStrictEqual([
|
||||
"width",
|
||||
"height",
|
||||
"batch_size",
|
||||
"upscale_method",
|
||||
"LatentUpscale width",
|
||||
"LatentUpscale height",
|
||||
"crop",
|
||||
"filename_prefix",
|
||||
]);
|
||||
});
|
||||
test("adds output for external links when converting to group", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const img = ez.EmptyLatentImage();
|
||||
let decode = ez.VAEDecode(...img.outputs);
|
||||
const preview1 = ez.PreviewImage(...decode.outputs);
|
||||
const preview2 = ez.PreviewImage(...decode.outputs);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [img, decode, preview1]);
|
||||
|
||||
// Ensure we have an output connected to the 2nd preview node
|
||||
expect(group.outputs.length).toBe(1);
|
||||
expect(group.outputs[0].connections.length).toBe(1);
|
||||
expect(group.outputs[0].connections[0].targetNode.id).toBe(preview2.id);
|
||||
|
||||
// Convert back and ensure bothe previews are still connected
|
||||
group.menu["Convert to nodes"].call();
|
||||
decode = graph.find(decode);
|
||||
expect(decode.outputs[0].connections.length).toBe(2);
|
||||
expect(decode.outputs[0].connections[0].targetNode.id).toBe(preview1.id);
|
||||
expect(decode.outputs[0].connections[1].targetNode.id).toBe(preview2.id);
|
||||
});
|
||||
test("adds output for external links when converting to group when nodes are not in execution order", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const sampler = ez.KSampler();
|
||||
const ckpt = ez.CheckpointLoaderSimple();
|
||||
const empty = ez.EmptyLatentImage();
|
||||
const pos = ez.CLIPTextEncode(ckpt.outputs.CLIP, { text: "positive" });
|
||||
const neg = ez.CLIPTextEncode(ckpt.outputs.CLIP, { text: "negative" });
|
||||
const decode1 = ez.VAEDecode(sampler.outputs.LATENT, ckpt.outputs.VAE);
|
||||
const save = ez.SaveImage(decode1.outputs.IMAGE);
|
||||
ckpt.outputs.MODEL.connectTo(sampler.inputs.model);
|
||||
pos.outputs.CONDITIONING.connectTo(sampler.inputs.positive);
|
||||
neg.outputs.CONDITIONING.connectTo(sampler.inputs.negative);
|
||||
empty.outputs.LATENT.connectTo(sampler.inputs.latent_image);
|
||||
|
||||
const encode = ez.VAEEncode(decode1.outputs.IMAGE);
|
||||
const vae = ez.VAELoader();
|
||||
const decode2 = ez.VAEDecode(encode.outputs.LATENT, vae.outputs.VAE);
|
||||
const preview = ez.PreviewImage(decode2.outputs.IMAGE);
|
||||
vae.outputs.VAE.connectTo(encode.inputs.vae);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [vae, decode1, encode, sampler]);
|
||||
|
||||
expect(group.outputs.length).toBe(3);
|
||||
expect(group.outputs[0].output.name).toBe("VAE");
|
||||
expect(group.outputs[0].output.type).toBe("VAE");
|
||||
expect(group.outputs[1].output.name).toBe("IMAGE");
|
||||
expect(group.outputs[1].output.type).toBe("IMAGE");
|
||||
expect(group.outputs[2].output.name).toBe("LATENT");
|
||||
expect(group.outputs[2].output.type).toBe("LATENT");
|
||||
|
||||
expect(group.outputs[0].connections.length).toBe(1);
|
||||
expect(group.outputs[0].connections[0].targetNode.id).toBe(decode2.id);
|
||||
expect(group.outputs[0].connections[0].targetInput.index).toBe(1);
|
||||
|
||||
expect(group.outputs[1].connections.length).toBe(1);
|
||||
expect(group.outputs[1].connections[0].targetNode.id).toBe(save.id);
|
||||
expect(group.outputs[1].connections[0].targetInput.index).toBe(0);
|
||||
|
||||
expect(group.outputs[2].connections.length).toBe(1);
|
||||
expect(group.outputs[2].connections[0].targetNode.id).toBe(decode2.id);
|
||||
expect(group.outputs[2].connections[0].targetInput.index).toBe(0);
|
||||
|
||||
expect((await graph.toPrompt()).output).toEqual({
|
||||
...getOutput({ 1: ckpt.id, 2: pos.id, 3: neg.id, 4: empty.id, 5: sampler.id, 6: decode1.id, 7: save.id }),
|
||||
[vae.id]: { inputs: { vae_name: "vae1.safetensors" }, class_type: vae.node.type },
|
||||
[encode.id]: { inputs: { pixels: ["6", 0], vae: [vae.id + "", 0] }, class_type: encode.node.type },
|
||||
[decode2.id]: { inputs: { samples: [encode.id + "", 0], vae: [vae.id + "", 0] }, class_type: decode2.node.type },
|
||||
[preview.id]: { inputs: { images: [decode2.id + "", 0] }, class_type: preview.node.type },
|
||||
});
|
||||
});
|
||||
test("works with IMAGEUPLOAD widget", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const img = ez.LoadImage();
|
||||
const preview1 = ez.PreviewImage(img.outputs[0]);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [img, preview1]);
|
||||
const widget = group.widgets["upload"];
|
||||
expect(widget).toBeTruthy();
|
||||
expect(widget.widget.type).toBe("button");
|
||||
});
|
||||
test("internal primitive populates widgets for all linked inputs", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const img = ez.LoadImage();
|
||||
const scale1 = ez.ImageScale(img.outputs[0]);
|
||||
const scale2 = ez.ImageScale(img.outputs[0]);
|
||||
ez.PreviewImage(scale1.outputs[0]);
|
||||
ez.PreviewImage(scale2.outputs[0]);
|
||||
|
||||
scale1.widgets.width.convertToInput();
|
||||
scale2.widgets.height.convertToInput();
|
||||
|
||||
const primitive = ez.PrimitiveNode();
|
||||
primitive.outputs[0].connectTo(scale1.inputs.width);
|
||||
primitive.outputs[0].connectTo(scale2.inputs.height);
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [img, primitive, scale1, scale2]);
|
||||
group.widgets.value.value = 100;
|
||||
expect((await graph.toPrompt()).output).toEqual({
|
||||
1: {
|
||||
inputs: { image: img.widgets.image.value, upload: "image" },
|
||||
class_type: "LoadImage",
|
||||
},
|
||||
2: {
|
||||
inputs: { upscale_method: "nearest-exact", width: 100, height: 512, crop: "disabled", image: ["1", 0] },
|
||||
class_type: "ImageScale",
|
||||
},
|
||||
3: {
|
||||
inputs: { upscale_method: "nearest-exact", width: 512, height: 100, crop: "disabled", image: ["1", 0] },
|
||||
class_type: "ImageScale",
|
||||
},
|
||||
4: { inputs: { images: ["2", 0] }, class_type: "PreviewImage" },
|
||||
5: { inputs: { images: ["3", 0] }, class_type: "PreviewImage" },
|
||||
});
|
||||
});
|
||||
test("primitive control widgets values are copied on convert", async () => {
|
||||
const { ez, graph, app } = await start();
|
||||
const sampler = ez.KSampler();
|
||||
sampler.widgets.seed.convertToInput();
|
||||
sampler.widgets.sampler_name.convertToInput();
|
||||
|
||||
let p1 = ez.PrimitiveNode();
|
||||
let p2 = ez.PrimitiveNode();
|
||||
p1.outputs[0].connectTo(sampler.inputs.seed);
|
||||
p2.outputs[0].connectTo(sampler.inputs.sampler_name);
|
||||
|
||||
p1.widgets.control_after_generate.value = "increment";
|
||||
p2.widgets.control_after_generate.value = "decrement";
|
||||
p2.widgets.control_filter_list.value = "/.*/";
|
||||
|
||||
p2.node.title = "p2";
|
||||
|
||||
const group = await convertToGroup(app, graph, "test", [sampler, p1, p2]);
|
||||
expect(group.widgets.control_after_generate.value).toBe("increment");
|
||||
expect(group.widgets["p2 control_after_generate"].value).toBe("decrement");
|
||||
expect(group.widgets["p2 control_filter_list"].value).toBe("/.*/");
|
||||
|
||||
group.widgets.control_after_generate.value = "fixed";
|
||||
group.widgets["p2 control_after_generate"].value = "randomize";
|
||||
group.widgets["p2 control_filter_list"].value = "/.+/";
|
||||
|
||||
group.menu["Convert to nodes"].call();
|
||||
p1 = graph.find(p1);
|
||||
p2 = graph.find(p2);
|
||||
|
||||
expect(p1.widgets.control_after_generate.value).toBe("fixed");
|
||||
expect(p2.widgets.control_after_generate.value).toBe("randomize");
|
||||
expect(p2.widgets.control_filter_list.value).toBe("/.+/");
|
||||
});
|
||||
});
|
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue