{
"4": {
"inputs": {
"ckpt_name": "stableDiffusion35_medium.safetensors"
},
"class_type": "CheckpointLoaderSimple",
"_meta": {
"title": "Load Checkpoint"
}
},
"6": {
"inputs": {
"text": "<%= it.prompt %>",
"clip": [
"11",
0
]
},
"class_type": "CLIPTextEncode",
"_meta": {
"title": "CLIP Text Encode (Prompt)"
}
},
"8": {
"inputs": {
"samples": [
"294",
0
],
"vae": [
"4",
2
]
},
"class_type": "VAEDecode",
"_meta": {
"title": "VAE Decode"
}
},
"11": {
"inputs": {
"clip_name1": "long_clip/ViT-L-14-TEXT-detail-improved-hiT-GmP-HF.safetensors",
"clip_name2": "clip_g.safetensors",
"clip_name3": "t5/google_t5-v1_1-xxl_encoderonly-fp8_e4m3fn.safetensors"
},
"class_type": "TripleCLIPLoader",
"_meta": {
"title": "TripleCLIPLoader"
}
},
"13": {
"inputs": {
"shift": 3,
"model": [
"4",
0
]
},
"class_type": "ModelSamplingSD3",
"_meta": {
"title": "ModelSamplingSD3"
}
},
"67": {
"inputs": {
"conditioning": [
"71",
0
]
},
"class_type": "ConditioningZeroOut",
"_meta": {
"title": "ConditioningZeroOut"
}
},
"68": {
"inputs": {
"start": 0.1,
"end": 1,
"conditioning": [
"67",
0
]
},
"class_type": "ConditioningSetTimestepRange",
"_meta": {
"title": "ConditioningSetTimestepRange"
}
},
"69": {
"inputs": {
"conditioning_1": [
"68",
0
],
"conditioning_2": [
"70",
0
]
},
"class_type": "ConditioningCombine",
"_meta": {
"title": "Conditioning (Combine)"
}
},
"70": {
"inputs": {
"start": 0,
"end": 0.1,
"conditioning": [
"71",
0
]
},
"class_type": "ConditioningSetTimestepRange",
"_meta": {
"title": "ConditioningSetTimestepRange"
}
},
"71": {
"inputs": {
"text": "<%= it.negativePrompt ||'' %>",
"clip": [
"11",
0
]
},
"class_type": "CLIPTextEncode",
"_meta": {
"title": "CLIP Text Encode (Prompt)"
}
},
"294": {
"inputs": {
"seed": <%= it.seed || it.randomSeed() %>,
"steps": <%= it.steps || 20 %>,
"cfg": <%= it.cfg || 2 %>,
"sampler_name": "<%= it.sampler || "euler" %>",
"scheduler": "<%= it.scheduler || 'beta' %>",
"denoise": 1,
"model": [
"13",
0
],
"positive": [
"6",
0
],
"negative": [
"69",
0
],
"latent_image": [
"302",
0
]
},
"class_type": "KSampler",
"_meta": {
"title": "KSampler"
}
},
"301": {
"inputs": {
"filename_prefix": "SD3.5_Medium_improved_",
"images": [
"325",
0
]
},
"class_type": "SaveImage",
"_meta": {
"title": "Save Image"
}
},
"302": {
"inputs": {
"batch_size": <%= it.batchSize || 1 %>,
"seed": <%= it.seed || it.randomSeed() %>
},
"class_type": "RandomEmptyLatentImageFromPresetsSDXL",
"_meta": {
"title": "Random Empty Latent Image From Presets (SDXL)"
}
},
"304": {
"inputs": {
"model_name": "sam_vit_h_4b8939.pth",
"device_mode": "CPU"
},
"class_type": "SAMLoader",
"_meta": {
"title": "SAMLoader (Impact)"
}
},
"305": {
"inputs": {
"max_faces": 10,
"face": true,
"mouth": false,
"left_eyebrow": false,
"left_eye": false,
"left_pupil": false,
"right_eyebrow": false,
"right_eye": false,
"right_pupil": false
},
"class_type": "MediaPipeFaceMeshDetectorProvider //Inspire",
"_meta": {
"title": "MediaPipeFaceMesh Detector Provider"
}
},
"306": {
"inputs": {
"model_name": "segm/person_yolov8m-seg.pt"
},
"class_type": "UltralyticsDetectorProvider",
"_meta": {
"title": "UltralyticsDetectorProvider"
}
},
"308": {
"inputs": {
"model_name": "bbox/hand_yolov8s.pt"
},
"class_type": "UltralyticsDetectorProvider",
"_meta": {
"title": "UltralyticsDetectorProvider"
}
},
"309": {
"inputs": {
"images": [
"324",
5
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"310": {
"inputs": {
"images": [
"324",
2
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"311": {
"inputs": {
"images": [
"324",
1
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"312": {
"inputs": {
"images": [
"324",
0
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"314": {
"inputs": {
"images": [
"325",
0
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"315": {
"inputs": {
"images": [
"325",
1
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"316": {
"inputs": {
"images": [
"325",
2
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"317": {
"inputs": {
"images": [
"325",
5
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"321": {
"inputs": {
"images": [
"8",
0
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
},
"324": {
"inputs": {
"guide_size": 512,
"guide_size_for": true,
"max_size": 1024,
"seed": <%= it.seed || it.randomSeed() %>,
"steps": <%= it.steps || 20 %>,
"cfg": <%= it.cfg || 2 %>,
"sampler_name": "<%= it.sampler || "euler" %>",
"scheduler": "<%= it.scheduler || 'beta' %>",
"denoise": 0.5,
"feather": 5,
"noise_mask": true,
"force_inpaint": true,
"bbox_threshold": 0.5,
"bbox_dilation": 10,
"bbox_crop_factor": 3,
"sam_detection_hint": "center-1",
"sam_dilation": 0,
"sam_threshold": 0.93,
"sam_bbox_expansion": 0,
"sam_mask_hint_threshold": 0.7,
"sam_mask_hint_use_negative": "False",
"drop_size": 10,
"wildcard": "",
"cycle": 1,
"inpaint_model": false,
"noise_mask_feather": 20,
"tiled_encode": false,
"tiled_decode": false,
"image": [
"8",
0
],
"model": [
"13",
0
],
"clip": [
"11",
0
],
"vae": [
"4",
2
],
"positive": [
"6",
0
],
"negative": [
"69",
0
],
"bbox_detector": [
"305",
0
],
"sam_model_opt": [
"304",
0
],
"segm_detector_opt": [
"305",
1
]
},
"class_type": "FaceDetailer",
"_meta": {
"title": "FaceDetailer"
}
},
"325": {
"inputs": {
"guide_size": 512,
"guide_size_for": true,
"max_size": 1024,
"seed": <%= it.seed || it.randomSeed() %>,
"steps": <%= it.steps || 20 %>,
"cfg": <%= it.cfg || 2 %>,
"sampler_name": "<%= it.sampler || "euler" %>",
"scheduler": "<%= it.scheduler || 'beta' %>",
"denoise": 0.5,
"feather": 5,
"noise_mask": true,
"force_inpaint": true,
"bbox_threshold": 0.5,
"bbox_dilation": 10,
"bbox_crop_factor": 3,
"sam_detection_hint": "center-1",
"sam_dilation": 0,
"sam_threshold": 0.93,
"sam_bbox_expansion": 0,
"sam_mask_hint_threshold": 0.7,
"sam_mask_hint_use_negative": "False",
"drop_size": 10,
"wildcard": "",
"cycle": 1,
"inpaint_model": false,
"noise_mask_feather": 20,
"tiled_encode": false,
"tiled_decode": false,
"image": [
"324",
0
],
"model": [
"13",
0
],
"clip": [
"11",
0
],
"vae": [
"4",
2
],
"positive": [
"6",
0
],
"negative": [
"69",
0
],
"bbox_detector": [
"308",
0
],
"sam_model_opt": [
"304",
0
],
"segm_detector_opt": [
"306",
1
]
},
"class_type": "FaceDetailer",
"_meta": {
"title": "FaceDetailer"
}
},
"save_image_websocket_node": {
"inputs": {
"images": [
"325",
0
]
},
"class_type": "SaveImageWebsocket",
"_meta": {
"title": "SaveImageWebsocket"
}
}
}