1-4 same promp different seed. Seed: #2: 795486351928825 #3: 795486351928826 #4: 795486351928827 {"5": {"inputs": {"samples": ["9", 1], "vae": ["9", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "7": {"inputs": {"base_ckpt_name": "ponyRealism_V22MainVAE.safetensors", "base_clip_skip": -2, "refiner_ckpt_name": "None", "refiner_clip_skip": -2, "positive_ascore": 6.0, "negative_ascore": 2.0, "vae_name": "Baked VAE", "positive": "score_9, 1girl, solo, depuss, outdoors, nude, anus, pink anus, pussy, plump labia, big pussy, cleft of venus, sitting, , presenting pussy,pussy, anus, masturbating, fisting pussy, small breasts, smile, furry female, cute, bottomless, slightly chubby, detailed fur, shirt, long sleeves, jacket, white shirt, pink jacket, Rebecca Cunningham, bear girl, brown fur, brown hair, back view, pink anus, orgasm face, , perfect face, , picnic blanket", "negative": "monochrome, 2d, text, artist signature, logo, nipples, bad anatomy, ugly face, blurry face, deformaed face, ", "token_normalization": "none", "weight_interpretation": "comfy", "empty_latent_width": 1024, "empty_latent_height": 1024, "batch_size": 4, "lora_stack": ["11", 0]}, "class_type": "Eff. Loader SDXL", "_meta": {"title": "Eff. Loader SDXL"}}, "9": {"inputs": {"noise_seed": 795486351928825, "steps": 45, "cfg": 7.5, "sampler_name": "euler_ancestral", "scheduler": "normal", "start_at_step": 0, "refine_at_step": -1, "preview_method": "auto", "vae_decode": "true", "sdxl_tuple": ["7", 0], "latent_image": ["7", 1], "optional_vae": ["7", 2]}, "class_type": "KSampler SDXL (Eff.)", "_meta": {"title": "KSampler SDXL (Eff.)"}}, "11": {"inputs": {"input_mode": "simple", "lora_count": 4, "lora_name_1": "rebecca_cunningham.safetensors", "lora_wt_1": 1.0, "model_str_1": 1.0, "clip_str_1": 1.0, "lora_name_2": "spo_sdxl_10ep_4k-data_lora_webui.safetensors", "lora_wt_2": 1.0, "model_str_2": 1.0, "clip_str_2": 1.0, "lora_name_3": "StS_PonyXL_Detail_Slider_v1.4_iteration_3.safetensors", "lora_wt_3": 0.55, "model_str_3": 1.0, "clip_str_3": 1.0, "lora_name_4": "None", "lora_wt_4": 0.35000000000000003, "model_str_4": 1.0, "clip_str_4": 1.0, "lora_name_5": "None", "lora_wt_5": 1.0, "model_str_5": 1.0, "clip_str_5": 1.0, "lora_name_6": "None", "lora_wt_6": 1.0, "model_str_6": 1.0, "clip_str_6": 1.0, "lora_name_7": "None", "lora_wt_7": 1.0, "model_str_7": 1.0, "clip_str_7": 1.0, "lora_name_8": "None", "lora_wt_8": 1.0, "model_str_8": 1.0, "clip_str_8": 1.0, "lora_name_9": "None", "lora_wt_9": 1.0, "model_str_9": 1.0, "clip_str_9": 1.0, "lora_name_10": "None", "lora_wt_10": 1.0, "model_str_10": 1.0, "clip_str_10": 1.0, "lora_name_11": "None", "lora_wt_11": 1.0, "model_str_11": 1.0, "clip_str_11": 1.0, "lora_name_12": "None", "lora_wt_12": 1.0, "model_str_12": 1.0, "clip_str_12": 1.0, "lora_name_13": "None", "lora_wt_13": 1.0, "model_str_13": 1.0, "clip_str_13": 1.0, "lora_name_14": "None", "lora_wt_14": 1.0, "model_str_14": 1.0, "clip_str_14": 1.0, "lora_name_15": "None", "lora_wt_15": 1.0, "model_str_15": 1.0, "clip_str_15": 1.0, "lora_name_16": "None", "lora_wt_16": 1.0, "model_str_16": 1.0, "clip_str_16": 1.0, "lora_name_17": "None", "lora_wt_17": 1.0, "model_str_17": 1.0, "clip_str_17": 1.0, "lora_name_18": "None", "lora_wt_18": 1.0, "model_str_18": 1.0, "clip_str_18": 1.0, "lora_name_19": "None", "lora_wt_19": 1.0, "model_str_19": 1.0, "clip_str_19": 1.0, "lora_name_20": "None", "lora_wt_20": 1.0, "model_str_20": 1.0, "clip_str_20": 1.0, "lora_name_21": "None", "lora_wt_21": 1.0, "model_str_21": 1.0, "clip_str_21": 1.0, "lora_name_22": "None", "lora_wt_22": 1.0, "model_str_22": 1.0, "clip_str_22": 1.0, "lora_name_23": "None", "lora_wt_23": 1.0, "model_str_23": 1.0, "clip_str_23": 1.0, "lora_name_24": "None", "lora_wt_24": 1.0, "model_str_24": 1.0, "clip_str_24": 1.0, "lora_name_25": "None", "lora_wt_25": 1.0, "model_str_25": 1.0, "clip_str_25": 1.0, "lora_name_26": "None", "lora_wt_26": 1.0, "model_str_26": 1.0, "clip_str_26": 1.0, "lora_name_27": "None", "lora_wt_27": 1.0, "model_str_27": 1.0, "clip_str_27": 1.0, "lora_name_28": "None", "lora_wt_28": 1.0, "model_str_28": 1.0, "clip_str_28": 1.0, "lora_name_29": "None", "lora_wt_29": 1.0, "model_str_29": 1.0, "clip_str_29": 1.0, "lora_name_30": "None", "lora_wt_30": 1.0, "model_str_30": 1.0, "clip_str_30": 1.0, "lora_name_31": "None", "lora_wt_31": 1.0, "model_str_31": 1.0, "clip_str_31": 1.0, "lora_name_32": "None", "lora_wt_32": 1.0, "model_str_32": 1.0, "clip_str_32": 1.0, "lora_name_33": "None", "lora_wt_33": 1.0, "model_str_33": 1.0, "clip_str_33": 1.0, "lora_name_34": "None", "lora_wt_34": 1.0, "model_str_34": 1.0, "clip_str_34": 1.0, "lora_name_35": "None", "lora_wt_35": 1.0, "model_str_35": 1.0, "clip_str_35": 1.0, "lora_name_36": "None", "lora_wt_36": 1.0, "model_str_36": 1.0, "clip_str_36": 1.0, "lora_name_37": "None", "lora_wt_37": 1.0, "model_str_37": 1.0, "clip_str_37": 1.0, "lora_name_38": "None", "lora_wt_38": 1.0, "model_str_38": 1.0, "clip_str_38": 1.0, "lora_name_39": "None", "lora_wt_39": 1.0, "model_str_39": 1.0, "clip_str_39": 1.0, "lora_name_40": "None", "lora_wt_40": 1.0, "model_str_40": 1.0, "clip_str_40": 1.0, "lora_name_41": "None", "lora_wt_41": 1.0, "model_str_41": 1.0, "clip_str_41": 1.0, "lora_name_42": "None", "lora_wt_42": 1.0, "model_str_42": 1.0, "clip_str_42": 1.0, "lora_name_43": "None", "lora_wt_43": 1.0, "model_str_43": 1.0, "clip_str_43": 1.0, "lora_name_44": "None", "lora_wt_44": 1.0, "model_str_44": 1.0, "clip_str_44": 1.0, "lora_name_45": "None", "lora_wt_45": 1.0, "model_str_45": 1.0, "clip_str_45": 1.0, "lora_name_46": "None", "lora_wt_46": 1.0, "model_str_46": 1.0, "clip_str_46": 1.0, "lora_name_47": "None", "lora_wt_47": 1.0, "model_str_47": 1.0, "clip_str_47": 1.0, "lora_name_48": "None", "lora_wt_48": 1.0, "model_str_48": 1.0, "clip_str_48": 1.0, "lora_name_49": "None", "lora_wt_49": 1.0, "model_str_49": 1.0, "clip_str_49": 1.0}, "class_type": "LoRA Stacker", "_meta": {"title": "LoRA Stacker"}}, "12": {"inputs": {"filename_prefix": "ComfyUI", "images": ["13", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "13": {"inputs": {"upscale_model": ["14", 0], "image": ["5", 0]}, "class_type": "ImageUpscaleWithModel", "_meta": {"title": "Upscale Image (using Model)"}}, "14": {"inputs": {"model_name": "4xlsdirplus_v1.pt"}, "class_type": "UpscaleModelLoader", "_meta": {"title": "Load Upscale Model"}}} 5 Seeds: Seed: #6: 735762599481352 #7: 735762599481353 {"5": {"inputs": {"samples": ["9", 1], "vae": ["9", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "7": {"inputs": {"base_ckpt_name": "ponyRealism_V22MainVAE.safetensors", "base_clip_skip": -2, "refiner_ckpt_name": "None", "refiner_clip_skip": -2, "positive_ascore": 6.0, "negative_ascore": 2.0, "vae_name": "Baked VAE", "positive": "score_9, 1girl, solo, depuss, outdoors, nude, anus, pink anus, pussy, plump labia, innie pussy, cleft of venus, presenting pussy,pussy, anus, spread pussy, presenting pussy, sitting, spreading legs, , feet, small breasts, smile, furry female, cute, bottomless, slightly chubby, detailed fur, shirt, long sleeves, jacket, white shirt, pink jacket, Rebecca Cunningham, bear girl, brown fur, brown hair, front view, spreading pussy, gaping pussy, big anus, pink anus, orgasm face, pubic hair, perfect face", "negative": "monochrome, 2d, text, artist signature, logo, nipples, lower angle, ugly face, blurry face, deformaed face, ", "token_normalization": "none", "weight_interpretation": "comfy", "empty_latent_width": 1024, "empty_latent_height": 1024, "batch_size": 10, "lora_stack": ["11", 0]}, "class_type": "Eff. Loader SDXL", "_meta": {"title": "Eff. Loader SDXL"}}, "9": {"inputs": {"noise_seed": 735762599481352, "steps": 45, "cfg": 7.5, "sampler_name": "euler_ancestral", "scheduler": "normal", "start_at_step": 0, "refine_at_step": -1, "preview_method": "auto", "vae_decode": "true", "sdxl_tuple": ["7", 0], "latent_image": ["7", 1], "optional_vae": ["7", 2]}, "class_type": "KSampler SDXL (Eff.)", "_meta": {"title": "KSampler SDXL (Eff.)"}}, "11": {"inputs": {"input_mode": "simple", "lora_count": 4, "lora_name_1": "rebecca_cunningham.safetensors", "lora_wt_1": 1.0, "model_str_1": 1.0, "clip_str_1": 1.0, "lora_name_2": "spo_sdxl_10ep_4k-data_lora_webui.safetensors", "lora_wt_2": 1.0, "model_str_2": 1.0, "clip_str_2": 1.0, "lora_name_3": "StS_PonyXL_Detail_Slider_v1.4_iteration_3.safetensors", "lora_wt_3": 0.55, "model_str_3": 1.0, "clip_str_3": 1.0, "lora_name_4": "None", "lora_wt_4": 0.35000000000000003, "model_str_4": 1.0, "clip_str_4": 1.0, "lora_name_5": "None", "lora_wt_5": 1.0, "model_str_5": 1.0, "clip_str_5": 1.0, "lora_name_6": "None", "lora_wt_6": 1.0, "model_str_6": 1.0, "clip_str_6": 1.0, "lora_name_7": "None", "lora_wt_7": 1.0, "model_str_7": 1.0, "clip_str_7": 1.0, "lora_name_8": "None", "lora_wt_8": 1.0, "model_str_8": 1.0, "clip_str_8": 1.0, "lora_name_9": "None", "lora_wt_9": 1.0, "model_str_9": 1.0, "clip_str_9": 1.0, "lora_name_10": "None", "lora_wt_10": 1.0, "model_str_10": 1.0, "clip_str_10": 1.0, "lora_name_11": "None", "lora_wt_11": 1.0, "model_str_11": 1.0, "clip_str_11": 1.0, "lora_name_12": "None", "lora_wt_12": 1.0, "model_str_12": 1.0, "clip_str_12": 1.0, "lora_name_13": "None", "lora_wt_13": 1.0, "model_str_13": 1.0, "clip_str_13": 1.0, "lora_name_14": "None", "lora_wt_14": 1.0, "model_str_14": 1.0, "clip_str_14": 1.0, "lora_name_15": "None", "lora_wt_15": 1.0, "model_str_15": 1.0, "clip_str_15": 1.0, "lora_name_16": "None", "lora_wt_16": 1.0, "model_str_16": 1.0, "clip_str_16": 1.0, "lora_name_17": "None", "lora_wt_17": 1.0, "model_str_17": 1.0, "clip_str_17": 1.0, "lora_name_18": "None", "lora_wt_18": 1.0, "model_str_18": 1.0, "clip_str_18": 1.0, "lora_name_19": "None", "lora_wt_19": 1.0, "model_str_19": 1.0, "clip_str_19": 1.0, "lora_name_20": "None", "lora_wt_20": 1.0, "model_str_20": 1.0, "clip_str_20": 1.0, "lora_name_21": "None", "lora_wt_21": 1.0, "model_str_21": 1.0, "clip_str_21": 1.0, "lora_name_22": "None", "lora_wt_22": 1.0, "model_str_22": 1.0, "clip_str_22": 1.0, "lora_name_23": "None", "lora_wt_23": 1.0, "model_str_23": 1.0, "clip_str_23": 1.0, "lora_name_24": "None", "lora_wt_24": 1.0, "model_str_24": 1.0, "clip_str_24": 1.0, "lora_name_25": "None", "lora_wt_25": 1.0, "model_str_25": 1.0, "clip_str_25": 1.0, "lora_name_26": "None", "lora_wt_26": 1.0, "model_str_26": 1.0, "clip_str_26": 1.0, "lora_name_27": "None", "lora_wt_27": 1.0, "model_str_27": 1.0, "clip_str_27": 1.0, "lora_name_28": "None", "lora_wt_28": 1.0, "model_str_28": 1.0, "clip_str_28": 1.0, "lora_name_29": "None", "lora_wt_29": 1.0, "model_str_29": 1.0, "clip_str_29": 1.0, "lora_name_30": "None", "lora_wt_30": 1.0, "model_str_30": 1.0, "clip_str_30": 1.0, "lora_name_31": "None", "lora_wt_31": 1.0, "model_str_31": 1.0, "clip_str_31": 1.0, "lora_name_32": "None", "lora_wt_32": 1.0, "model_str_32": 1.0, "clip_str_32": 1.0, "lora_name_33": "None", "lora_wt_33": 1.0, "model_str_33": 1.0, "clip_str_33": 1.0, "lora_name_34": "None", "lora_wt_34": 1.0, "model_str_34": 1.0, "clip_str_34": 1.0, "lora_name_35": "None", "lora_wt_35": 1.0, "model_str_35": 1.0, "clip_str_35": 1.0, "lora_name_36": "None", "lora_wt_36": 1.0, "model_str_36": 1.0, "clip_str_36": 1.0, "lora_name_37": "None", "lora_wt_37": 1.0, "model_str_37": 1.0, "clip_str_37": 1.0, "lora_name_38": "None", "lora_wt_38": 1.0, "model_str_38": 1.0, "clip_str_38": 1.0, "lora_name_39": "None", "lora_wt_39": 1.0, "model_str_39": 1.0, "clip_str_39": 1.0, "lora_name_40": "None", "lora_wt_40": 1.0, "model_str_40": 1.0, "clip_str_40": 1.0, "lora_name_41": "None", "lora_wt_41": 1.0, "model_str_41": 1.0, "clip_str_41": 1.0, "lora_name_42": "None", "lora_wt_42": 1.0, "model_str_42": 1.0, "clip_str_42": 1.0, "lora_name_43": "None", "lora_wt_43": 1.0, "model_str_43": 1.0, "clip_str_43": 1.0, "lora_name_44": "None", "lora_wt_44": 1.0, "model_str_44": 1.0, "clip_str_44": 1.0, "lora_name_45": "None", "lora_wt_45": 1.0, "model_str_45": 1.0, "clip_str_45": 1.0, "lora_name_46": "None", "lora_wt_46": 1.0, "model_str_46": 1.0, "clip_str_46": 1.0, "lora_name_47": "None", "lora_wt_47": 1.0, "model_str_47": 1.0, "clip_str_47": 1.0, "lora_name_48": "None", "lora_wt_48": 1.0, "model_str_48": 1.0, "clip_str_48": 1.0, "lora_name_49": "None", "lora_wt_49": 1.0, "model_str_49": 1.0, "clip_str_49": 1.0}, "class_type": "LoRA Stacker", "_meta": {"title": "LoRA Stacker"}}, "12": {"inputs": {"filename_prefix": "ComfyUI", "images": ["13", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "13": {"inputs": {"upscale_model": ["14", 0], "image": ["5", 0]}, "class_type": "ImageUpscaleWithModel", "_meta": {"title": "Upscale Image (using Model)"}}, "14": {"inputs": {"model_name": "4xlsdirplus_v1.pt"}, "class_type": "UpscaleModelLoader", "_meta": {"title": "Load Upscale Model"}}} 8 Seed: #9: 516978970540668 #10:516978970540669 #11: 516978970540670 {"5": {"inputs": {"samples": ["9", 1], "vae": ["9", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "7": {"inputs": {"base_ckpt_name": "ponyRealism_V22MainVAE.safetensors", "base_clip_skip": -2, "refiner_ckpt_name": "None", "refiner_clip_skip": -2, "positive_ascore": 6.0, "negative_ascore": 2.0, "vae_name": "Baked VAE", "positive": "score_9, 1girl, solo, depuss, outdoors, nude, anus, pink anus, pussy, plump labia, big pussy, cleft of venus, bent over, presenting butt, presenting pussy,pussy, anus, standing, , small breasts, smile, furry female, cute, bottomless, slightly chubby, detailed fur, shirt, long sleeves, jacket, white shirt, pink jacket, Rebecca Cunningham, bear girl, brown fur, brown hair, back view, spreading pussy, pussy, big anus, pink anus, orgasm face, pubic hair, perfect face, hand on pussy, ", "negative": "monochrome, 2d, text, artist signature, logo, nipples, bad anatomy, ugly face, blurry face, deformaed face, ", "token_normalization": "none", "weight_interpretation": "comfy", "empty_latent_width": 1024, "empty_latent_height": 1024, "batch_size": 4, "lora_stack": ["11", 0]}, "class_type": "Eff. Loader SDXL", "_meta": {"title": "Eff. Loader SDXL"}}, "9": {"inputs": {"noise_seed": 516978970540668, "steps": 45, "cfg": 7.5, "sampler_name": "euler_ancestral", "scheduler": "normal", "start_at_step": 0, "refine_at_step": -1, "preview_method": "auto", "vae_decode": "true", "sdxl_tuple": ["7", 0], "latent_image": ["7", 1], "optional_vae": ["7", 2]}, "class_type": "KSampler SDXL (Eff.)", "_meta": {"title": "KSampler SDXL (Eff.)"}}, "11": {"inputs": {"input_mode": "simple", "lora_count": 4, "lora_name_1": "rebecca_cunningham.safetensors", "lora_wt_1": 1.0, "model_str_1": 1.0, "clip_str_1": 1.0, "lora_name_2": "spo_sdxl_10ep_4k-data_lora_webui.safetensors", "lora_wt_2": 1.0, "model_str_2": 1.0, "clip_str_2": 1.0, "lora_name_3": "StS_PonyXL_Detail_Slider_v1.4_iteration_3.safetensors", "lora_wt_3": 0.55, "model_str_3": 1.0, "clip_str_3": 1.0, "lora_name_4": "None", "lora_wt_4": 0.35000000000000003, "model_str_4": 1.0, "clip_str_4": 1.0, "lora_name_5": "None", "lora_wt_5": 1.0, "model_str_5": 1.0, "clip_str_5": 1.0, "lora_name_6": "None", "lora_wt_6": 1.0, "model_str_6": 1.0, "clip_str_6": 1.0, "lora_name_7": "None", "lora_wt_7": 1.0, "model_str_7": 1.0, "clip_str_7": 1.0, "lora_name_8": "None", "lora_wt_8": 1.0, "model_str_8": 1.0, "clip_str_8": 1.0, "lora_name_9": "None", "lora_wt_9": 1.0, "model_str_9": 1.0, "clip_str_9": 1.0, "lora_name_10": "None", "lora_wt_10": 1.0, "model_str_10": 1.0, "clip_str_10": 1.0, "lora_name_11": "None", "lora_wt_11": 1.0, "model_str_11": 1.0, "clip_str_11": 1.0, "lora_name_12": "None", "lora_wt_12": 1.0, "model_str_12": 1.0, "clip_str_12": 1.0, "lora_name_13": "None", "lora_wt_13": 1.0, "model_str_13": 1.0, "clip_str_13": 1.0, "lora_name_14": "None", "lora_wt_14": 1.0, "model_str_14": 1.0, "clip_str_14": 1.0, "lora_name_15": "None", "lora_wt_15": 1.0, "model_str_15": 1.0, "clip_str_15": 1.0, "lora_name_16": "None", "lora_wt_16": 1.0, "model_str_16": 1.0, "clip_str_16": 1.0, "lora_name_17": "None", "lora_wt_17": 1.0, "model_str_17": 1.0, "clip_str_17": 1.0, "lora_name_18": "None", "lora_wt_18": 1.0, "model_str_18": 1.0, "clip_str_18": 1.0, "lora_name_19": "None", "lora_wt_19": 1.0, "model_str_19": 1.0, "clip_str_19": 1.0, "lora_name_20": "None", "lora_wt_20": 1.0, "model_str_20": 1.0, "clip_str_20": 1.0, "lora_name_21": "None", "lora_wt_21": 1.0, "model_str_21": 1.0, "clip_str_21": 1.0, "lora_name_22": "None", "lora_wt_22": 1.0, "model_str_22": 1.0, "clip_str_22": 1.0, "lora_name_23": "None", "lora_wt_23": 1.0, "model_str_23": 1.0, "clip_str_23": 1.0, "lora_name_24": "None", "lora_wt_24": 1.0, "model_str_24": 1.0, "clip_str_24": 1.0, "lora_name_25": "None", "lora_wt_25": 1.0, "model_str_25": 1.0, "clip_str_25": 1.0, "lora_name_26": "None", "lora_wt_26": 1.0, "model_str_26": 1.0, "clip_str_26": 1.0, "lora_name_27": "None", "lora_wt_27": 1.0, "model_str_27": 1.0, "clip_str_27": 1.0, "lora_name_28": "None", "lora_wt_28": 1.0, "model_str_28": 1.0, "clip_str_28": 1.0, "lora_name_29": "None", "lora_wt_29": 1.0, "model_str_29": 1.0, "clip_str_29": 1.0, "lora_name_30": "None", "lora_wt_30": 1.0, "model_str_30": 1.0, "clip_str_30": 1.0, "lora_name_31": "None", "lora_wt_31": 1.0, "model_str_31": 1.0, "clip_str_31": 1.0, "lora_name_32": "None", "lora_wt_32": 1.0, "model_str_32": 1.0, "clip_str_32": 1.0, "lora_name_33": "None", "lora_wt_33": 1.0, "model_str_33": 1.0, "clip_str_33": 1.0, "lora_name_34": "None", "lora_wt_34": 1.0, "model_str_34": 1.0, "clip_str_34": 1.0, "lora_name_35": "None", "lora_wt_35": 1.0, "model_str_35": 1.0, "clip_str_35": 1.0, "lora_name_36": "None", "lora_wt_36": 1.0, "model_str_36": 1.0, "clip_str_36": 1.0, "lora_name_37": "None", "lora_wt_37": 1.0, "model_str_37": 1.0, "clip_str_37": 1.0, "lora_name_38": "None", "lora_wt_38": 1.0, "model_str_38": 1.0, "clip_str_38": 1.0, "lora_name_39": "None", "lora_wt_39": 1.0, "model_str_39": 1.0, "clip_str_39": 1.0, "lora_name_40": "None", "lora_wt_40": 1.0, "model_str_40": 1.0, "clip_str_40": 1.0, "lora_name_41": "None", "lora_wt_41": 1.0, "model_str_41": 1.0, "clip_str_41": 1.0, "lora_name_42": "None", "lora_wt_42": 1.0, "model_str_42": 1.0, "clip_str_42": 1.0, "lora_name_43": "None", "lora_wt_43": 1.0, "model_str_43": 1.0, "clip_str_43": 1.0, "lora_name_44": "None", "lora_wt_44": 1.0, "model_str_44": 1.0, "clip_str_44": 1.0, "lora_name_45": "None", "lora_wt_45": 1.0, "model_str_45": 1.0, "clip_str_45": 1.0, "lora_name_46": "None", "lora_wt_46": 1.0, "model_str_46": 1.0, "clip_str_46": 1.0, "lora_name_47": "None", "lora_wt_47": 1.0, "model_str_47": 1.0, "clip_str_47": 1.0, "lora_name_48": "None", "lora_wt_48": 1.0, "model_str_48": 1.0, "clip_str_48": 1.0, "lora_name_49": "None", "lora_wt_49": 1.0, "model_str_49": 1.0, "clip_str_49": 1.0}, "class_type": "LoRA Stacker", "_meta": {"title": "LoRA Stacker"}}, "12": {"inputs": {"filename_prefix": "ComfyUI", "images": ["13", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "13": {"inputs": {"upscale_model": ["14", 0], "image": ["5", 0]}, "class_type": "ImageUpscaleWithModel", "_meta": {"title": "Upscale Image (using Model)"}}, "14": {"inputs": {"model_name": "4xlsdirplus_v1.pt"}, "class_type": "UpscaleModelLoader", "_meta": {"title": "Load Upscale Model"}}} 12 Seeds #13:947084713396633 #14:947084713396634 #15: 947084713396635 {"5": {"inputs": {"samples": ["9", 1], "vae": ["9", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "7": {"inputs": {"base_ckpt_name": "ponyRealism_V22MainVAE.safetensors", "base_clip_skip": -2, "refiner_ckpt_name": "None", "refiner_clip_skip": -2, "positive_ascore": 6.0, "negative_ascore": 2.0, "vae_name": "Baked VAE", "positive": "score_9, 1girl, solo, depuss, outdoors, nude, anus, pink anus, pussy, plump labia, big pussy, cleft of venus, squatting, presenting butt, presenting pussy,pussy, anus, standing, , small breasts, smile, furry female, cute, bottomless, slightly chubby, detailed fur, shirt, long sleeves, jacket, white shirt, pink jacket, Rebecca Cunningham, bear girl, brown fur, brown hair, back view, pink anus, orgasm face, , perfect face, , ", "negative": "monochrome, 2d, text, artist signature, logo, nipples, bad anatomy, ugly face, blurry face, deformaed face, ", "token_normalization": "none", "weight_interpretation": "comfy", "empty_latent_width": 1024, "empty_latent_height": 1024, "batch_size": 4, "lora_stack": ["11", 0]}, "class_type": "Eff. Loader SDXL", "_meta": {"title": "Eff. Loader SDXL"}}, "9": {"inputs": {"noise_seed": 947084713396632, "steps": 45, "cfg": 7.5, "sampler_name": "euler_ancestral", "scheduler": "normal", "start_at_step": 0, "refine_at_step": -1, "preview_method": "auto", "vae_decode": "true", "sdxl_tuple": ["7", 0], "latent_image": ["7", 1], "optional_vae": ["7", 2]}, "class_type": "KSampler SDXL (Eff.)", "_meta": {"title": "KSampler SDXL (Eff.)"}}, "11": {"inputs": {"input_mode": "simple", "lora_count": 4, "lora_name_1": "rebecca_cunningham.safetensors", "lora_wt_1": 1.0, "model_str_1": 1.0, "clip_str_1": 1.0, "lora_name_2": "spo_sdxl_10ep_4k-data_lora_webui.safetensors", "lora_wt_2": 1.0, "model_str_2": 1.0, "clip_str_2": 1.0, "lora_name_3": "StS_PonyXL_Detail_Slider_v1.4_iteration_3.safetensors", "lora_wt_3": 0.55, "model_str_3": 1.0, "clip_str_3": 1.0, "lora_name_4": "None", "lora_wt_4": 0.35000000000000003, "model_str_4": 1.0, "clip_str_4": 1.0, "lora_name_5": "None", "lora_wt_5": 1.0, "model_str_5": 1.0, "clip_str_5": 1.0, "lora_name_6": "None", "lora_wt_6": 1.0, "model_str_6": 1.0, "clip_str_6": 1.0, "lora_name_7": "None", "lora_wt_7": 1.0, "model_str_7": 1.0, "clip_str_7": 1.0, "lora_name_8": "None", "lora_wt_8": 1.0, "model_str_8": 1.0, "clip_str_8": 1.0, "lora_name_9": "None", "lora_wt_9": 1.0, "model_str_9": 1.0, "clip_str_9": 1.0, "lora_name_10": "None", "lora_wt_10": 1.0, "model_str_10": 1.0, "clip_str_10": 1.0, "lora_name_11": "None", "lora_wt_11": 1.0, "model_str_11": 1.0, "clip_str_11": 1.0, "lora_name_12": "None", "lora_wt_12": 1.0, "model_str_12": 1.0, "clip_str_12": 1.0, "lora_name_13": "None", "lora_wt_13": 1.0, "model_str_13": 1.0, "clip_str_13": 1.0, "lora_name_14": "None", "lora_wt_14": 1.0, "model_str_14": 1.0, "clip_str_14": 1.0, "lora_name_15": "None", "lora_wt_15": 1.0, "model_str_15": 1.0, "clip_str_15": 1.0, "lora_name_16": "None", "lora_wt_16": 1.0, "model_str_16": 1.0, "clip_str_16": 1.0, "lora_name_17": "None", "lora_wt_17": 1.0, "model_str_17": 1.0, "clip_str_17": 1.0, "lora_name_18": "None", "lora_wt_18": 1.0, "model_str_18": 1.0, "clip_str_18": 1.0, "lora_name_19": "None", "lora_wt_19": 1.0, "model_str_19": 1.0, "clip_str_19": 1.0, "lora_name_20": "None", "lora_wt_20": 1.0, "model_str_20": 1.0, "clip_str_20": 1.0, "lora_name_21": "None", "lora_wt_21": 1.0, "model_str_21": 1.0, "clip_str_21": 1.0, "lora_name_22": "None", "lora_wt_22": 1.0, "model_str_22": 1.0, "clip_str_22": 1.0, "lora_name_23": "None", "lora_wt_23": 1.0, "model_str_23": 1.0, "clip_str_23": 1.0, "lora_name_24": "None", "lora_wt_24": 1.0, "model_str_24": 1.0, "clip_str_24": 1.0, "lora_name_25": "None", "lora_wt_25": 1.0, "model_str_25": 1.0, "clip_str_25": 1.0, "lora_name_26": "None", "lora_wt_26": 1.0, "model_str_26": 1.0, "clip_str_26": 1.0, "lora_name_27": "None", "lora_wt_27": 1.0, "model_str_27": 1.0, "clip_str_27": 1.0, "lora_name_28": "None", "lora_wt_28": 1.0, "model_str_28": 1.0, "clip_str_28": 1.0, "lora_name_29": "None", "lora_wt_29": 1.0, "model_str_29": 1.0, "clip_str_29": 1.0, "lora_name_30": "None", "lora_wt_30": 1.0, "model_str_30": 1.0, "clip_str_30": 1.0, "lora_name_31": "None", "lora_wt_31": 1.0, "model_str_31": 1.0, "clip_str_31": 1.0, "lora_name_32": "None", "lora_wt_32": 1.0, "model_str_32": 1.0, "clip_str_32": 1.0, "lora_name_33": "None", "lora_wt_33": 1.0, "model_str_33": 1.0, "clip_str_33": 1.0, "lora_name_34": "None", "lora_wt_34": 1.0, "model_str_34": 1.0, "clip_str_34": 1.0, "lora_name_35": "None", "lora_wt_35": 1.0, "model_str_35": 1.0, "clip_str_35": 1.0, "lora_name_36": "None", "lora_wt_36": 1.0, "model_str_36": 1.0, "clip_str_36": 1.0, "lora_name_37": "None", "lora_wt_37": 1.0, "model_str_37": 1.0, "clip_str_37": 1.0, "lora_name_38": "None", "lora_wt_38": 1.0, "model_str_38": 1.0, "clip_str_38": 1.0, "lora_name_39": "None", "lora_wt_39": 1.0, "model_str_39": 1.0, "clip_str_39": 1.0, "lora_name_40": "None", "lora_wt_40": 1.0, "model_str_40": 1.0, "clip_str_40": 1.0, "lora_name_41": "None", "lora_wt_41": 1.0, "model_str_41": 1.0, "clip_str_41": 1.0, "lora_name_42": "None", "lora_wt_42": 1.0, "model_str_42": 1.0, "clip_str_42": 1.0, "lora_name_43": "None", "lora_wt_43": 1.0, "model_str_43": 1.0, "clip_str_43": 1.0, "lora_name_44": "None", "lora_wt_44": 1.0, "model_str_44": 1.0, "clip_str_44": 1.0, "lora_name_45": "None", "lora_wt_45": 1.0, "model_str_45": 1.0, "clip_str_45": 1.0, "lora_name_46": "None", "lora_wt_46": 1.0, "model_str_46": 1.0, "clip_str_46": 1.0, "lora_name_47": "None", "lora_wt_47": 1.0, "model_str_47": 1.0, "clip_str_47": 1.0, "lora_name_48": "None", "lora_wt_48": 1.0, "model_str_48": 1.0, "clip_str_48": 1.0, "lora_name_49": "None", "lora_wt_49": 1.0, "model_str_49": 1.0, "clip_str_49": 1.0}, "class_type": "LoRA Stacker", "_meta": {"title": "LoRA Stacker"}}, "12": {"inputs": {"filename_prefix": "ComfyUI", "images": ["13", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "13": {"inputs": {"upscale_model": ["14", 0], "image": ["5", 0]}, "class_type": "ImageUpscaleWithModel", "_meta": {"title": "Upscale Image (using Model)"}}, "14": {"inputs": {"model_name": "4xlsdirplus_v1.pt"}, "class_type": "UpscaleModelLoader", "_meta": {"title": "Load Upscale Model"}}}