Stable Diffusion XL lets you create better, bigger pictures, with faces that look more real. To add to the customizability, it also supports swapping between SDXL models and SD 1. 0 with ComfyUI. For me I have an 8 gig vram, trying sdxl in auto1111 just tells me insufficient memory if it even loads the model and when running with --medvram image generation takes a whole lot of time, comfi ui is just better in that case for me. 6, as it makes inpainted part fit better into the overall image. SDXL-Inpainting is designed to make image editing smarter and more efficient. 9 through Python 3. This in-depth tutorial will guide you to set up repositories, prepare datasets, optimize training parameters, and leverage techniques like LoRA and inpainting to achieve photorealistic results. Stable Diffusion XL (SDXL) is a brand-new model with unprecedented performance. Use global_inpaint_harmonious when you want to set the inpainting denoising strength high. (optional) download Fixed SDXL 0. SDXL’s current out-of-the-box output falls short of a finely-tuned Stable Diffusion model. This model is available on Mage. zoupishness7 • 11 days ago. - The 2. 0) "Latent noise mask" does exactly what it says. 0 has been out for just a few weeks now, and already we're getting even more. x. 0-mid; controlnet-depth-sdxl-1. 0. It is a more flexible and accurate way to control the image generation process. 🚀LCM update brings SDXL and SSD-1B to the game 🎮Stable Diffusion XL (SDXL) Inpainting. For SD1. 5. Model type: Diffusion-based text-to-image generative model. I don’t think “if you’re too newb to figure it out try again later” is a. Discord can help give 1:1 troubleshooting (a lot of active contributors) - InvokeAI's WebUI interface is gorgeous and much more responsive than AUTOMATIC1111's. 400. 9 vae (335 MB) and copy it into ComfyUI/models/vae (instead of using the VAE that's embedded in SDXL 1. 6M runs stable-diffusion-inpainting Fill in masked parts of images with Stable Diffusion Updated 4 months, 2 weeks ago 15. 5 model. 0 with both the base and refiner checkpoints. 5 is in where you'll be spending your energy. Both are capable at txt2img, img2img, inpainting, upscaling, and so on. Nexustar. The settings I used are. 0-mid; We also encourage you to train custom ControlNets; we provide a training script for this. 0 with its predecessor, Stable Diffusion 2. The only important thing is that for optimal performance the resolution should be set to 1024x1024 or other resolutions with the same amount of pixels but a different aspect ratio. Alternatively, upgrade your transformers and accelerate package to latest. For the rest of methods (original, latent noise, latent nothing) 0,8 which is. SDXL 1. Positive Prompt; Negative Prompt; That’s it! There are a few more complex SDXL. @landmann If you are referring to small changes, than it is most likely due to the encoding/decoding step of the pipeline. The workflow also has TXT2IMG, IMG2IMG, up to 3x IP Adapter, 2x Revision, predefined (and editable) styles, optional up-scaling, Control Net Canny, Control Net Depth, Lora, selection of recommended SDXL resolutions, adjusting input images to the closest SDXL resolution, etc. (especially with SDXL which can work in plenty of aspect ratios). ago. Take the image out to a 1. This model can follow a two-stage model process (though each model can also be used alone); the base model generates an image, and a refiner model takes that image and further enhances its details and quality. SDXL v1. This is a fine-tuned. The Stable Diffusion model can also be applied to inpainting which lets you edit specific parts of an image by providing a mask and a text prompt using Stable Diffusion. 3 GB! Place it in the ComfyUI models\unet folder. Posted by u/Edzomatic - 9 votes and 3 commentsI'll teach you what you need to know about Inpainting in this Stable diffusion tutorial. Now I'm scared. 0 in ComfyUI I've come across three different methods that seem to be commonly used: Base Model with Latent Noise Mask, Base Model using InPaint VAE Encode and using the UNET "diffusion_pytorch" InPaint specific model from Hugging Face. 5. Inpainting using SDXL base kinda sucks (see diffusers issue #4392), and requires workarounds like hybrid (SD 1. SDXL + Inpainting + ControlNet pipeline . It's a transformative tool for. You can Load these images in ComfyUI to get the full workflow. go to img2img, choose batch, dropdown refiner, use the folder in 1 as input and the folder in 2 as output. This is why we also expose a CLI argument namely --pretrained_vae_model_name_or_path that lets you specify the location of a better VAE (such as this one). 75 for large changes. Home - Xcel Painting 317-652-7004. Stable Diffusion XL (SDXL) Inpainting. The purpose of DreamShaper has always been to make "a better Stable Diffusion", a model capable of doing everything on its own, to weave dreams. View more examples . I found some pretty strange render times (total VRAM 10240 MB, total RAM 32677 MB). Tout d'abord, SDXL 1. 3-inpainting File Name realisticVisionV20_v13-inpainting. LaMa: Resolution-robust Large Mask Inpainting with Fourier Convolutions (Apache-2. 6 billion, compared with 0. I was excited to learn SD to enhance my workflow. Stable Diffusion Inpainting, Stable Diffusion XL (SDXL) Inpainting, and Kandinsky 2. Right now I inpaint without controlnet, I just create the mask, let's say with clipseg, and just send in the mask for inpainting and it works okay (not super reliably, maybe 50% of the time it does something decent). Then that input image was used in the new Instruct-pix2pix tab ( now available in Auto1111 by adding an extension and model), with the text in each caption entered in the prompt field, using the default settings, except steps was changed. 0 is a drastic improvement to Stable Diffusion 2. Disclaimer: This post has been copied from lllyasviel's github post. Google Colab updated as well for ComfyUI and SDXL 1. You blur as a preprocessing instead of downsampling like you do with tile. I've found that the refiner tends to. Select "Add Difference". pytorch image-generation diffusers sdxl Updated Oct 25, 2023; Python. This model is available on Mage. py script pre-computes text embeddings and the VAE encodings and keeps them in memory. Stability AI on Huggingface: Here you can find all official SDXL models We might release a beta version of this feature before 3. SDXL 1. 3. v1 models are 1. 0 with its. Captain_MC_Henriques. 5 inpainting model but had no luck so far. • 19 days ago. • 13 days ago. The SDXL model allows users to effortlessly generate images based on text prompts. 0 license) Roman Suvorov, Elizaveta Logacheva, Anton Mashikhin, Anastasia Remizova, Arsenii Ashukha, Aleksei Silvestrov, Naejin Kong, Harshith Goka, Kiwoong Park, Victor LempitskyPlongeons dans les détails. On the right, the results of inpainting with SDXL 1. Outpainting with SDXL. 8 Comments. How to Achieve Perfect Results with SDXL Inpainting: Techniques and Strategies A step-by-step guide to maximizing the potential of the SDXL inpaint model for image transformation. The abstract from the paper is: We present SDXL, a latent diffusion model for text-to. Stable Diffusion XL (SDXL) Inpainting. Rest assured that we are working with Huggingface to address these issues with the Diffusers package. Inpainting. You can draw a mask or scribble to guide how it should inpaint/outpaint. Safety filter far less intrusive due to safe model design. Settings for Stable Diffusion SDXL Automatic1111 Controlnet. Nov 16,. Work on hands and bad anatomy with mask blur 4, inpaint at full resolution, masked content: original, 32 padding, denoise 0. Stable Inpainting also upgraded to v2. Navigate to the 'Inpainting' section within the 'Inpaint Anything' tab and click the Get prompt from: txt2img (or img2img) button. 5 models. I've been searching around online but cant find any info. r/StableDiffusion. 5, and their main competitor: MidJourney. v1. With SDXL (and, of course, DreamShaper XL 😉) just released, I think the " swiss knife " type of model is closer then ever. I was happy to finally have an SDXL based inpainting model, but I noticed an issue with it: the inpainted area gets a discoloration with a random intensity. Better human anatomy. on 1. In researching InPainting using SDXL 1. Inpainting is limited to what is essentially already there, you can't change the whole setup or pose or stuff like that with Inpainting (well, I guess theoretically you could, but the results would likely be crap). But everyone posting images of SDXL are just posting trash that looks like a bad day on launch day of midjourney v4 back in November. While it can do regular txt2img and img2img, it really shines when filling in missing regions. 0. ago • Edited 6 mo. 5 and 2. To get the best inpainting results you should therefore resize your Bounding Box to the smallest area that contains your mask and. r/StableDiffusion. With this, you can get the faces you've grown to love, while benefiting from the highly detailed SDXL model. Features beyond image generation. Navigate to the 'Inpainting' section within the 'Inpaint Anything' tab and click the Get prompt from: txt2img (or img2img) button. I loved invokeAI and used it exclusively until a git pull broke it beyond reparation. Then ported it into Photoshop for further finishing a slight gradient layer to enhance the warm to cool lighting. This looks sexy, thanks. Stable Diffusion XL Inpainting is a state-of-the-art model that represents the pinnacle of image inpainting technology. Learn how to fix any Stable diffusion generated image through inpain. Stable Diffusion XL (SDXL) Inpainting. It copys the weights of neural network blocks into a "locked" copy and a "trainable" copy. Being the control freak that I am, I took the base refiner image into Automatic111 and inpainted the eyes and lips. I think it's possible to create similar patch model for SD 1. 1 is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask. 0 的过程,包括下载必要的模型以及如何将它们安装到. InvokeAI is an excellent implementation that has become very popular for its stability and ease of use for outpainting and inpainting edits. Inpainting has been used to reconstruct deteriorated images, eliminating imperfections like cracks, scratches, disfigured limbs, dust spots, or red-eye effects from AI-generated images. 3 on Civitai for download . 5. It is common to see extra or missing limbs. 0 Base Model + Refiner. Stable Diffusion XL (SDXL) was proposed in SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis by Dustin Podell, Zion English, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas Müller, Joe Penna, and Robin Rombach. I made a textual inversion for the artist Jeff Delgado. New Inpainting Model. Carmel, IN 46032. 5 inpainting model but had no luck so far. The SDXL inpainting model cannot be found in the model download list. Because of its larger size, the base model itself. 9, the most advanced version to date, offers a remarkable enhancement in image and composition detail compared to its predecessor. In the top Preview Bridge, right click and mask the area you want to inpaint. I had interpreted it, since he mentioned it in his question, that he was trying to use controlnet with inpainting which would cause problems naturally with sdxl. rachelwearsshoes • 5 mo. 1. 22. Задача inpainting намного сложнее, чем стандартная генерация, потому что модели нужно научиться генерировать. That image is really good btw 👌. Using IMG2IMG Automatic 1111 tool in SDXL. r/StableDiffusion. There is a “Pad Image for Outpainting” node to automatically pad the image for outpainting while creating the proper mask. Developed by: Stability AI. This GUI is similar to the Huggingface demo, but you won't have to wait. Two models are available. It was developed by researchers. Alternatively, upgrade your transformers and accelerate package to latest. SDXL will require even more RAM to generate larger images. 0. More information can be found here. 5). Stable Diffusion XL (SDXL) is a powerful text-to-image generation model that iterates on the previous Stable Diffusion models in three key ways: the UNet is 3x larger and SDXL combines a second text encoder (OpenCLIP ViT-bigG/14) with the original text encoder to significantly increase the number of parameters. 3. What Is Inpainting? Inpainting is a technique used in Stable Diffusion image editing to restore and edit missing or damaged portions of pictures. Unveiling the Magic of Artistic Creations with Stable Diffusion XL Inpainting. 🎁 Benefits: 🥇 Be among the first to test SDXL-beta with Automatic1111! ⚡ Experience lightning-fast and cost-effective. 0) using your own dataset with the Segmind training module. This model is a specialized variant of the renowned Stable Diffusion series, designed to seamlessly fill in and reconstruct parts of images with astonishing accuracy and detail. No ControlNet, No ADetailer, No LoRAs, No inpainting, No editing, No face restoring, Not Even Hires Fix!! (and obviously no spaghetti nightmare). It is a much larger model. What is the SDXL Inpainting Desktop Client and Why Does It Matter? Imagine a desktop application that uses AI to paint parts of an image masked by you. Notes . Settings for Stable Diffusion SDXL Automatic1111 Controlnet. use increment or fixed. make a folder in img2img. Code. Go to checkpoint merger and drop sd1. 0. 17:38 How to use inpainting with SDXL with ComfyUI. MultiControlnet with inpainting in diffusers doesn't exist as of now. 10 Stable Diffusion extensions for next-level creativity. Inpainting. So for example, if I have a 512x768 image, with a full body and smaller / zoomed out face, I inpaint the face, but change the res to 1024x1536, and it gives better detail and definition to the area I am inpainting. For some reason the inpainting black is still there but invisible. Commercial. For example: 896x1152 or 1536x640 are good resolutions. One trick that was on here a few weeks ago to make an inpainting model from any other model based on SD1. You can use it with or without mask in lama cleaner. The developer posted these notes about the update: A big step-up from V1. The chart above evaluates user preference for SDXL (with and without refinement) over SDXL 0. I cant' confirm the Pixel Art XL lora works with other ones. jpg ^ --mask mask. Take the image out to a 1. 7. Stable Diffusion XL. In the center, the results of inpainting with Stable Diffusion 2. 1. Nov 17, 2023 4 min read. SDXL-Inpainting is designed to make image editing smarter and more efficient. 1, or Windows 8. Mataric. Read here for a list of tips for optimizing inference: Optimum-SDXL-Usage. 0 base and have lots of fun with it. Intelligent sampler defaults. The question is not whether people will run one or the other. They're the do-anything tools. That model architecture is big and heavy enough to accomplish that the. Our clients choose to work with us because they want quality craftsmanship. With a ControlNet model, you can provide an additional control image to condition and control Stable Diffusion generation. 5 inpainting model though if I'm not mistaken. 9 doesn't seem to work with less than 1024×1024, and so it uses around 8-10 gb vram even at the bare minimum for 1 image batch due to the model being loaded itself as well The max I can do on 24gb vram is 6 image batch of 1024×1024. We follow the original repository and provide basic inference scripts to sample from the models. This model can follow a two-stage model process (though each model can also be used alone); the base model generates an image, and a refiner model takes that image and further enhances its details and quality. For example my base image is 512x512. I'm not 100% because I haven't tested it myself, but I do believe you can use a higher noise ratio with ControlNet inpainting vs. There’s a ton of naming confusion here. Outpainting is the same thing as inpainting. Edit model card. Select Controlnet preprocessor "inpaint_only+lama". Model Description: This is a model that can be used to generate and modify images based on text prompts. By using a mask to pinpoint the areas that need enhancement and applying inpainting, you can effectively improve the visual quality of facial features while preserving the overall composition. ComfyUI shared workflows are also updated for SDXL 1. Try to add "pixel art" at the start of the prompt, and your style and the end, for example: "pixel art, a dinosaur on a forest, landscape, ghibli style". 20:57 How to use LoRAs with SDXL. The SDXL inpainting model cannot be found in the model download list. I cranked up the number of steps for faces, no idea if that. It offers a feathering option but it's generally not needed and you can actually get better results by simply increasing the grow_mask_by in the VAE Encode (for Inpainting) node. from diffusers import StableDiffusionControlNetInpaintPipeline controlnet = [ ControlNetModel. Check add differences and hit go. 5 pruned. 1 is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures. 0 ComfyUI workflows! Fancy something that in. png ^ --hint sketch. 5 and SD1. Enter the right KSample parameters. 5 my workflow used to be: 1- Img-Img upscale (this corrected a lot of details 2- Inpainting with controlnet (got decent results) 3- Controlnet tile for upscale 4- Upscale the image with upscalers This workflow doesn't work for SDXL, and I'd love to know what workflow. The inpainting model, which is saved in HuggingFace's cache and includes inpaint (case-insensitive) in its repo_id, will also be added to the Inpainting Model ID dropdown list. In this article, we’ll compare the results of SDXL 1. As before, it will allow you to mask sections of the. Applying inpainting to SDXL-generated images can be effective in fixing specific facial regions that lack detail or accuracy. 9 and ran it through ComfyUI. 5 I added the (masterpiece) and (best quality) modifiers to each prompt, and with SDXL I added the offset lora of . 0 (B1) Status (Updated: Nov 18, 2023): - Training Images: +2620 - Training Steps: +524k - Approximate percentage of. The abstract from the paper is: We present SDXL, a latent diffusion model for text-to-image synthesis. That model architecture is big and heavy enough to accomplish that the. 2. Outpainting just uses a normal model. 222 added a new inpaint preprocessor: inpaint_only+lama . 11. SDXL is a larger and more powerful version of Stable Diffusion v1. To use them, right click on your desired workflow, press "Download Linked File". Some of these features will be forthcoming releases from Stability. 5 models. 9 can be used for various applications, including films, television, music, instructional videos, and design and industrial use. 3. Stable Diffusion web UIのInpainting機能について Inpaintingとは? Inpainting(web UI内だと「inpaint」という表記になっています)は 画像の一部のみを修正するときに便利な機能 です。 自分で塗りつぶした部分だけに呪文を適用できるので、希望する部分だけを簡単に変更することができます。Welcome to the 🧨 diffusers organization! diffusers is the go-to library for state-of-the-art pretrained diffusion models for multi-modal generative AI. Strategies for optimizing the SDXL inpaint model for high quality outputs: Here, we'll discuss strategies and settings to help you get the most out of the SDXL inpaint model, ensuring high-quality and precise image outputs. VRAM settings. 0-base. Actions. • 2 mo. 0 in ComfyUI I've come across three different methods that seem to be commonly used: Base Model with Latent Noise Mask, Base Model using InPaint VAE Encode and using the UNET "diffusion_pytorch" InPaint specific model from Hugging Face. Exploring Alternative. Read More. Reply reply more replies. This is the same as Photoshop’s new generative fill function, but free. (up to 1024/1024), might be even higher for SDXL, your model becomes more flexible at running at random aspects ratios or even just set up your subject as a side part of a bigger image and so on. SDXL is a larger and more powerful version of Stable Diffusion v1. When using a Lora model, you're making a full image of that in whatever setup you want. 0 and Refiner 1. Stable Diffusion XL (SDXL), is the latest AI image generation model that can generate realistic faces, legible text within the images, and better image composition, all while using shorter and simpler prompts. * The result should best be in the resolution-space of SDXL (1024x1024). Say you inpaint an area, generate, download the image. PS内直接跑图,模型可自由控制!. . [SDXL LoRA] - "LucasArts Artstyle" - 90s PC Adventure game / Pixelart model - (I try not to pimp my own civitai content, but I. 以下. Stable Diffusion is a deep learning, text-to-image model released in 2022 based on diffusion techniques. Use the paintbrush tool to create a mask on the area you want to regenerate. This model can follow a two-stage model process (though each model can also be used alone); the base model generates an image, and a refiner model takes that image and further enhances its details and quality. Although it is not yet perfect (his own words), you can use it and have fun. I find the results interesting for comparison; hopefully others will too. 0 will be, hopefully it doesnt require a refiner model because dual model workflows are much more inflexible to work with. Clearly, SDXL 1. Updating ControlNet. Btw, I usually use an anime model to do the fixing, because they are trained with clearer outlined images for body parts (typical for manga, anime), and finish the pipeline with a realistic model for refining. 3 ; Always use the latest version of the workflow json file with the latest. x for ComfyUI. 0 Features: Shared VAE Load: the. 5-Inpainting) Set "B" to your model. Automatic1111 will NOT work with SDXL until it's been updated. Select Controlnet preprocessor "inpaint_only+lama". Researchers discover that Stable Diffusion v1 uses internal representations of 3D geometry when generating an image. Upload the image to the inpainting canvas. 0 is a new text-to-image model by Stability AI. 5 billion. Support for FreeU has been added and is included in the v4. 🚀Announcing stable-fast v0. Stable Diffusion XL. 5 (on civitai it shows you near the download button). 🚀LCM update brings SDXL and SSD-1B to the game 🎮Inpainting is not particularly good at inserting brand new subjects into an image, and if that’s your goal, you are better off image bashing or scribbling it in, or doing multiple inpainting passes (usually 3-4). Sample codes are below: # for depth conditioned controlnet python test_controlnet_inpaint_sd_xl_depth. 1. 📷 All of the flexibility of Stable Diffusion: SDXL is primed for complex image design workflows that include generation for text or base image, inpainting (with masks), outpainting, and more. The workflow also has TXT2IMG, IMG2IMG, up to 3x IP Adapter, 2x Revision, predefined (and editable) styles, optional up-scaling, Control Net Canny, Control Net Depth, Lora, selection of recommended SDXL resolutions, adjusting input images to the closest SDXL resolution, etc. (I have heard different opinions about the VAE not being necessary to be selected manually since it is baked in the model but still to make sure I use manual mode) 3) Then I write a prompt, set resolution of the image output at 1024. Custom nodes extension for ComfyUI, including a workflow to use SDXL 1. 4000 W. Here is a link for more information. Stable Diffusion XL specifically trained on Inpainting by huggingface. [2023/8/30] 🔥 Add an IP-Adapter with face image as prompt. With SD 1. SDXL typically produces higher resolution images than Stable Diffusion v1. 3) will revert to default SDXL model when trying to load non-SDXL model. 6. But neither the base model or the refiner is particularly good at generating images from images that noise has been added to (img2img generation), and the refiner even does a poor job doing an img2img render at 0. backafterdeleting. /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. 9 and Stable Diffusion 1. Use the brush tool in the Controlnet image panel to paint over the part of the image you want to change. As before, it will allow you to mask sections of the image you would like to let the model have another go at generating, letting you make changes and adjustments to the content or just having another go at a hand that doesn’t. Mask mode: Inpaint masked. The problem with it is that the inpainting is performed at the whole resolution image, which makes the model perform poorly on already upscaled images. To use ControlNet inpainting: It is best to use the same model that generates the image. sdxl sdxl lora sdxl inpainting comfyui. 78. 5. SDXL is a larger and more powerful version of Stable Diffusion v1. SDXL has an inpainting model, but I haven't found a way to merge it with other models yet. It may help to use the inpainting model, but not. 2 Inpainting are among the most popular models for inpainting. Inpainting Workflow for ComfyUI. It is primarily used to generate detailed images conditioned on text descriptions, though it can also be applied to other tasks such as inpainting, outpainting, and generating image-to-image translations guided by a text prompt. SDXL Unified Canvas Together with ControlNet and SDXL LoRAs, the Unified Canvas becomes a robust platform for unparalleled editing, generation, and manipulation. ComfyUI Master Tutorial - Stable Diffusion XL (SDXL) - Install On PC, Google Colab (Free) & RunPod, SDXL LoRA, SDXL InPainting. 5. After generating an image on the txt2img page, click Send to Inpaint to send the image to the Inpaint tab on the Img2img page. With SDXL (and, of course, DreamShaper XL 😉) just released, I think the " swiss knife " type of model is closer then ever. The SDXL Desktop client is a powerful UI for inpainting images using Stable Diffusion XL. Im curious if its possible to do a training on the 1. Most other inpainting/outpainting apps use Stable Diffusion's standard inpainting function, which has trouble filling in blank areas with things that make sense and fit visually with the rest of the image. 0. I trained a LoRA model of myself using the SDXL 1. SDXL has an inpainting model, but I haven't found a way to merge it with other models yet. The chart above evaluates user preference for SDXL (with and without refinement) over SDXL 0. It comes with some optimizations that bring the VRAM usage down to 7-9GB, depending on how large of an image you are working with. No constructure change has been. 5 model. Start Free Trial Upgrade Today. safetensors or diffusion_pytorch_model. diffusers/stable-diffusion-xl-1. 19k. Free Stable Diffusion inpainting. python inpaint. Is there something I'm missing about how to do what we used to call out painting for SDXL images?. Just like its predecessors, SDXL has the ability to generate image variations using image-to-image prompting, inpainting (reimagining of. The company says it represents a key step forward in its image generation models. It can be used in combination with Stable Diffusion, such as runwayml/stable-diffusion-v1-5. Making your own inpainting model is very simple: Go to Checkpoint Merger. controlnet doesn't work with SDXL yet so not possible. How to Achieve Perfect Results with SDXL Inpainting: Techniques and Strategies A step-by-step guide to maximizing the potential of the SDXL inpaint model for image transformation. g. An infinite zoom art is a visual art technique that creates an illusion of an infinite zoom-in or zoom-out on. This repository provides the implementation of StableDiffusionXLControlNetInpaintPipeline and. Moreover, SDXL has functionality that extends beyond just text-to-image prompting, including image-to-image prompting (inputting one image to get variations of that image), inpainting. Here's what I've found: When I pair the SDXL base with my LoRA on ComfyUI, things seem to click and work pretty well. Sped up SDXL generation from 4 mins to 25 seconds!A good place to start if you have no idea how any of this works is the: ComfyUI Basic Tutorial VN: All the art is made with ComfyUI. In our experiments, we found that SDXL yields good initial results without extensive hyperparameter tuning. zoupishness7 • 11 days ago.