From 4e8cb7e472a10e029b844431fcbf39cbc0cadd35 Mon Sep 17 00:00:00 2001 From: Abdullah Alfaraj <7842232+AbdullahAlfaraj@users.noreply.github.com> Date: Mon, 27 Nov 2023 20:34:48 +0300 Subject: [PATCH] add ip adapter custom workflow examples --- .../IPAdapter_simple_api.json | 167 ++++++++++++++++ .../IPAdapter_weighted_api.json | 180 ++++++++++++++++++ 2 files changed, 347 insertions(+) create mode 100644 typescripts/comfyui/native_workflows/IPAdapter_simple_api.json create mode 100644 typescripts/comfyui/native_workflows/IPAdapter_weighted_api.json diff --git a/typescripts/comfyui/native_workflows/IPAdapter_simple_api.json b/typescripts/comfyui/native_workflows/IPAdapter_simple_api.json new file mode 100644 index 0000000..eefb2de --- /dev/null +++ b/typescripts/comfyui/native_workflows/IPAdapter_simple_api.json @@ -0,0 +1,167 @@ +{ + "8": { + "inputs": { + "samples": [ + "50", + 0 + ], + "vae": [ + "32", + 0 + ] + }, + "class_type": "VAEDecode" + }, + "14": { + "inputs": { + "ckpt_name": "dreamshaper_8.safetensors" + }, + "class_type": "CheckpointLoaderSimple" + }, + "32": { + "inputs": { + "vae_name": "vae-ft-mse-840000-ema-pruned.safetensors" + }, + "class_type": "VAELoader" + }, + "50": { + "inputs": { + "add_noise": "enable", + "noise_seed": 538408362410054, + "steps": 30, + "cfg": 7, + "sampler_name": "dpmpp_2m", + "scheduler": "karras", + "start_at_step": 0, + "end_at_step": 10000, + "return_with_leftover_noise": "disable", + "model": [ + "210", + 0 + ], + "positive": [ + "211", + 0 + ], + "negative": [ + "212", + 0 + ], + "latent_image": [ + "95", + 0 + ] + }, + "class_type": "KSamplerAdvanced" + }, + "95": { + "inputs": { + "width": 512, + "height": 512, + "batch_size": 1 + }, + "class_type": "EmptyLatentImage" + }, + "192": { + "inputs": { + "images": [ + "8", + 0 + ] + }, + "class_type": "PreviewImage" + }, + "204": { + "inputs": { + "weight": 0.5, + "noise": 0.33, + "weight_type": "original", + "start_at": 0, + "end_at": 1, + "ipadapter": [ + "205", + 0 + ], + "clip_vision": [ + "206", + 0 + ], + "image": [ + "207", + 0 + ], + "model": [ + "14", + 0 + ] + }, + "class_type": "IPAdapterApply" + }, + "205": { + "inputs": { + "ipadapter_file": "ip-adapter-plus_sd15.bin" + }, + "class_type": "IPAdapterModelLoader" + }, + "206": { + "inputs": { + "clip_name": "model.safetensors" + }, + "class_type": "CLIPVisionLoader" + }, + "207": { + "inputs": { + "interpolation": "LANCZOS", + "crop_position": "top", + "sharpening": 0.15, + "image": [ + "209", + 0 + ] + }, + "class_type": "PrepImageForClipVision" + }, + "209": { + "inputs": { + "image": "ComfyUI_temp_cqoqp_00001_ (1).png", + "choose file to upload": "image" + }, + "class_type": "LoadImage" + }, + "210": { + "inputs": { + "block_number": 3, + "downscale_factor": 1.5, + "start_percent": 0, + "end_percent": 0.45, + "downscale_after_skip": true, + "downscale_method": "bicubic", + "upscale_method": "bicubic", + "model": [ + "204", + 0 + ] + }, + "class_type": "PatchModelAddDownscale" + }, + "211": { + "inputs": { + "text": "", + "clip": [ + "14", + 1 + ] + }, + "class_type": "CLIPTextEncode" + }, + "212": { + "inputs": { + "text": "blurry, low quality", + "clip": [ + "14", + 1 + ] + }, + "class_type": "CLIPTextEncode" + } +} \ No newline at end of file diff --git a/typescripts/comfyui/native_workflows/IPAdapter_weighted_api.json b/typescripts/comfyui/native_workflows/IPAdapter_weighted_api.json new file mode 100644 index 0000000..91bc53a --- /dev/null +++ b/typescripts/comfyui/native_workflows/IPAdapter_weighted_api.json @@ -0,0 +1,180 @@ +{ + "1": { + "inputs": { + "ckpt_name": "dreamshaper_8.safetensors" + }, + "class_type": "CheckpointLoaderSimple" + }, + "2": { + "inputs": { + "vae_name": "vae-ft-mse-840000-ema-pruned.safetensors" + }, + "class_type": "VAELoader" + }, + "3": { + "inputs": { + "ipadapter_file": "ip-adapter-plus_sd15.bin" + }, + "class_type": "IPAdapterModelLoader" + }, + "4": { + "inputs": { + "clip_name": "model.safetensors" + }, + "class_type": "CLIPVisionLoader" + }, + "6": { + "inputs": { + "image": "ComfyUI_temp_cqoqp_00001_ (1).png", + "choose file to upload": "image" + }, + "class_type": "LoadImage" + }, + "7": { + "inputs": { + "text": "beautiful renaissance girl, detailed", + "clip": [ + "1", + 1 + ] + }, + "class_type": "CLIPTextEncode" + }, + "8": { + "inputs": { + "text": "blurry, horror", + "clip": [ + "1", + 1 + ] + }, + "class_type": "CLIPTextEncode" + }, + "9": { + "inputs": { + "seed": 857545940756658, + "steps": 35, + "cfg": 5, + "sampler_name": "ddim", + "scheduler": "ddim_uniform", + "denoise": 1, + "model": [ + "13", + 0 + ], + "positive": [ + "7", + 0 + ], + "negative": [ + "8", + 0 + ], + "latent_image": [ + "10", + 0 + ] + }, + "class_type": "KSampler" + }, + "10": { + "inputs": { + "width": 512, + "height": 512, + "batch_size": 4 + }, + "class_type": "EmptyLatentImage" + }, + "11": { + "inputs": { + "samples": [ + "9", + 0 + ], + "vae": [ + "2", + 0 + ] + }, + "class_type": "VAEDecode" + }, + "12": { + "inputs": { + "filename_prefix": "IPAdapter", + "images": [ + "11", + 0 + ] + }, + "class_type": "SaveImage" + }, + "13": { + "inputs": { + "weight": 1, + "weight_type": "original", + "start_at": 0, + "end_at": 1, + "ipadapter": [ + "3", + 0 + ], + "embeds": [ + "14", + 0 + ], + "model": [ + "1", + 0 + ] + }, + "class_type": "IPAdapterApplyEncoded" + }, + "14": { + "inputs": { + "ipadapter_plus": true, + "noise": 0.31, + "weight_1": [ + "16", + 0 + ], + "weight_2": [ + "17", + 0 + ], + "weight_3": 1, + "weight_4": 1, + "clip_vision": [ + "4", + 0 + ], + "image_1": [ + "15", + 0 + ], + "image_2": [ + "6", + 0 + ] + }, + "class_type": "IPAdapterEncoder" + }, + "15": { + "inputs": { + "image": "animation-deku-my-hero-academia.jpg", + "choose file to upload": "image" + }, + "class_type": "LoadImage" + }, + "16": { + "inputs": { + "Value": 1 + }, + "class_type": "Float" + }, + "17": { + "inputs": { + "Value": 0.38 + }, + "class_type": "Float" + } +} \ No newline at end of file