PNG IHDR @ @ Na tEXtprompt {"3": {"inputs": {"seed": 357050682708314, "steps": 16, "cfg": 6.0, "sampler_name": "uni_pc", "scheduler": "normal", "denoise": 1.0, "model": ["15", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler"}, "4": {"inputs": {"config_name": "v1-inference_clip_skip_2.yaml", "ckpt_name": "anyloraCheckpoint_bakedvaeFtmseFp16NOT.safetensors"}, "class_type": "CheckpointLoader"}, "5": {"inputs": {"width": 576, "height": 576, "batch_size": 1}, "class_type": "EmptyLatentImage"}, "6": {"inputs": {"text": "green background, solo, blush, smile, long curly hair, shirt, green hair, long sleeves, 1boy, holding, hair between eyes, jewelry, jacket, white shirt, upper body, male focus, earrings, necklace, scarf, grin, black eyes, black jacket, grey eyes, fur trim, bandages, phone, piercing, cellphone, ear piercing, bandaid, smartphone, holding phone, bandaid on face", "clip": ["15", 1]}, "class_type": "CLIPTextEncode"}, "7": {"inputs": {"text": "embedding:easynegative embedding:badhands (hands), text, error, cropped, (worst quality:1.2), (low quality:1.2), normal quality, (jpeg artifacts:1.3), signature, watermark, username, blurry, artist name, monochrome, sketch, censorship, censor, (copyright:1.2), extra legs, (forehead mark) (depth of field) (emotionless) (penis) (pumpkin)", "clip": ["15", 1]}, "class_type": "CLIPTextEncode"}, "8": {"inputs": {"samples": ["3", 0], "vae": ["13", 0]}, "class_type": "VAEDecode"}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage"}, "10": {"inputs": {"strength": 1}, "class_type": "ControlNetApply"}, "11": {"inputs": {"image": "B (9).png", "choose file to upload": "image"}, "class_type": "LoadImage", "is_changed": "54fd798bf95e6f5ee7b6460ee564c41d6b9f01c8a82a46829e384c15d678c91f"}, "12": {"inputs": {"control_net_name": "control_openpose-fp16.safetensors"}, "class_type": "ControlNetLoader"}, "13": {"inputs": {"vae_name": "blessed2.vae.pt"}, "class_type": "VAELoader"}, "14": {"inputs": {"lora_name": "nami_pre_timeskip_offset.safetensors", "strength_model": 1, "strength_clip": 1}, "class_type": "LoraLoader"}, "15": {"inputs": {"lora_name": "simpleBackgrounds-10.safetensors", "strength_model": 1.0, "strength_clip": 1.0, "model": ["4", 0], "clip": ["4", 1]}, "class_type": "LoraLoader"}}xƆ vtEXtworkflow {"last_node_id": 15, "last_link_id": 37, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1210, 250], "size": {"0": 210, "1": 46}, "flags": {}, "order": 11, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 14}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "links": [9], "slot_index": 0}], "properties": {"Node name for S&R": "VAEDecode"}}, {"id": 13, "type": "VAELoader", "pos": [833, 484], "size": {"0": 315, "1": 58}, "flags": {}, "order": 0, "mode": 0, "outputs": [{"name": "VAE", "type": "VAE", "links": [14], "slot_index": 0}], "properties": {"Node name for S&R": "VAELoader"}, "widgets_values": ["blessed2.vae.pt"]}, {"id": 9, "type": "SaveImage", "pos": [1456, 248], "size": {"0": 210, "1": 250}, "flags": {}, "order": 12, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "properties": {}, "widgets_values": ["ComfyUI"]}, {"id": 5, "type": "EmptyLatentImage", "pos": [341, 618], "size": {"0": 315, "1": 106}, "flags": {}, "order": 1, "mode": 0, "outputs": [{"name": "LATENT", "type": "LATENT", "links": [2], "slot_index": 0}], "properties": {"Node name for S&R": "EmptyLatentImage"}, "widgets_values": [576, 576, 1]}, {"id": 14, "type": "LoraLoader", "pos": [-955, 801], "size": {"0": 315, "1": 126}, "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": null}, {"name": "clip", "type": "CLIP", "link": null}], "outputs": [{"name": "MODEL", "type": "MODEL", "links": [], "slot_index": 0}, {"name": "CLIP", "type": "CLIP", "links": [], "slot_index": 1}], "properties": {"Node name for S&R": "LoraLoader"}, "widgets_values": ["nami_pre_timeskip_offset.safetensors", 1, 1]}, {"id": 7, "type": "CLIPTextEncode", "pos": [200, 385], "size": {"0": 425.27801513671875, "1": 180.6060791015625}, "flags": {}, "order": 9, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 36}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "links": [16], "slot_index": 0}], "properties": {"Node name for S&R": "CLIPTextEncode"}, "widgets_values": ["embedding:easynegative embedding:badhands (hands), text, error, cropped, (worst quality:1.2), (low quality:1.2), normal quality, (jpeg artifacts:1.3), signature, watermark, username, blurry, artist name, monochrome, sketch, censorship, censor, (copyright:1.2), extra legs, (forehead mark) (depth of field) (emotionless) (penis) (pumpkin)"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [-26, -176], "size": {"0": 422.84503173828125, "1": 164.31304931640625}, "flags": {}, "order": 8, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 35}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "links": [24], "slot_index": 0}], "properties": {"Node name for S&R": "CLIPTextEncode"}, "widgets_values": ["green background, solo, blush, smile, long curly hair, shirt, green hair, long sleeves, 1boy, holding, hair between eyes, jewelry, jacket, white shirt, upper body, male focus, earrings, necklace, scarf, grin, black eyes, black jacket, grey eyes, fur trim, bandages, phone, piercing, cellphone, ear piercing, bandaid, smartphone, holding phone, bandaid on face"]}, {"id": 12, "type": "ControlNetLoader", "pos": [-26, 41], "size": {"0": 422, "1": 58}, "flags": {}, "order": 3, "mode": 0, "outputs": [{"name": "CONTROL_NET", "type": "CONTROL_NET", "links": [], "slot_index": 0}], "properties": {"Node name for S&R": "ControlNetLoader"}, "widgets_values": ["control_openpose-fp16.safetensors"]}, {"id": 11, "type": "LoadImage", "pos": [-25, 147], "size": {"0": 315, "1": 102}, "flags": {}, "order": 4, "mode": 0, "outputs": [{"name": "IMAGE", "type": "IMAGE", "links": [], "slot_index": 0}, {"name": "MASK", "type": "MASK", "links": null}], "properties": {"Node name for S&R": "LoadImage"}, "widgets_values": ["B (9).png", "image"]}, {"id": 3, "type": "KSampler", "pos": [747, 130], "size": {"0": 315, "1": 262}, "flags": {}, "order": 10, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 37}, {"name": "positive", "type": "CONDITIONING", "link": 24}, {"name": "negative", "type": "CONDITIONING", "link": 16}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "links": [7], "slot_index": 0}], "properties": {"Node name for S&R": "KSampler"}, "widgets_values": [357050682708314, "fixed", 16, 6, "uni_pc", "normal", 1]}, {"id": 10, "type": "ControlNetApply", "pos": [525, -153], "size": {"0": 317.4000244140625, "1": 98}, "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "conditioning", "type": "CONDITIONING", "link": null}, {"name": "control_net", "type": "CONTROL_NET", "link": null}, {"name": "image", "type": "IMAGE", "link": null}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "links": [], "slot_index": 0}], "properties": {"Node name for S&R": "ControlNetApply"}, "widgets_values": [1]}, {"id": 4, "type": "CheckpointLoader", "pos": [-1049, 391], "size": {"0": 315, "1": 122}, "flags": {}, "order": 6, "mode": 0, "outputs": [{"name": "MODEL", "type": "MODEL", "links": [30], "slot_index": 0}, {"name": "CLIP", "type": "CLIP", "links": [31], "slot_index": 1}, {"name": "VAE", "type": "VAE", "links": [], "slot_index": 2}], "properties": {"Node name for S&R": "CheckpointLoader"}, "widgets_values": ["v1-inference_clip_skip_2.yaml", "anyloraCheckpoint_bakedvaeFtmseFp16NOT.safetensors"]}, {"id": 15, "type": "LoraLoader", "pos": [-592, 383], "size": {"0": 315, "1": 126}, "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 30}, {"name": "clip", "type": "CLIP", "link": 31}], "outputs": [{"name": "MODEL", "type": "MODEL", "links": [37], "slot_index": 0}, {"name": "CLIP", "type": "CLIP", "links": [35, 36], "slot_index": 1}], "properties": {"Node name for S&R": "LoraLoader"}, "widgets_values": ["simpleBackgrounds-10.safetensors", 1, 1]}], "links": [[2, 5, 0, 3, 3, "LATENT"], [7, 3, 0, 8, 0, "LATENT"], [9, 8, 0, 9, 0, "IMAGE"], [14, 13, 0, 8, 1, "VAE"], [16, 7, 0, 3, 2, "CONDITIONING"], [24, 6, 0, 3, 1, "CONDITIONING"], [30, 4, 0, 15, 0, "MODEL"], [31, 4, 1, 15, 1, "CLIP"], [35, 15, 1, 6, 0, "CLIP"], [36, 15, 1, 7, 0, "CLIP"], [37, 15, 0, 3, 0, "MODEL"]], "groups": [], "config": {}, "extra": {}, "version": 0.4} IDATx^w%Q'ȬYo$A$a88;p!B$$BHHZyiwތlUeFȬ{q'N{i##"3#*ꑪ ~~uwI^ HE(D@ADP( "M,k(ʆ@ 3Q6Zkf˳)s+{;/;" +TAQä _xR/~Ζ^;cxI)jj)[8'}MAPPQ016͒**+B c0Rb A 8vRwkpȘ[WfZl(xx3_4G=sWqO'\|bw~%_Mo2["JP@%ч2sIcptyttϟ2ղ2yyO%]ŭV̦o=xԡΓPb^ETDTd92&P){U(œ1il$zRϐ%!%0G:PIBPse7ue23*x-*N$/]S0\:Fzǯ?h~dmv5y_}]tWq]Y]',_?!g U'*`cÁKD D"Y-jzlmoBTmZB@"DؐIyHs`+IBi -8|-{~,JW8x uT^K_rqRna-Y&CPH^\˶2ZBJ K2U +p) /*JoUU 6k &TlZKZbs04`/PO_n>8vwۣgU?5Ә ^|pŁ]'(>^W|m$a!/"UbV@DT$IF>uۛ|P[cRà;i&CL3`RXe$^}GR}wB" /NXx'bz8)1!$/A0l齂 r1YvhHa#U Ymh5ꍺ5FW U/|(YʵX֠lw:nq8ܺye(zKʟ/xy+8pEat: oW7i[ &x DURUԶf[3ss`"rb21%5 1, f&mXrc>[Zj&S[ŧvyދOo:'^2QX,T{$n[ u.IY?bP]T3)? NJ5.T7(x#o(lAH [f2SKMΩ,K$Vmjy\ӯ]қ?__~%\|