Stable diffusion upscale reddit

I'm sure this has been done to death, but here is a comparison of the different upscalers for some wants-to-be-photorealistic content. I just did a brand new install of Stability Matrix / Automatic1111, partly in hopes of fixing the following problem. It's only in the ui. When generating at higher resolutions than the model is trained at, the composition doesn’t get scaled, the composition of the additional space is more or less tiled, which is why you are more likely to get double torsos and other issues. 12 votes, 13 comments. I finally settled on the extras tab with the settings. Edit ui-config. 5x and 2x upscale option that is pretty neat. Do not send a massive file size into it or it will create so many tiles it will take an age to process. Random notes: - x4plus and 4x+ appear identical. Repeat until you're nice and clear. AnimeGenius is an anime upscaler that can turn pictures to a larger resolution with one click. Thanks for sharing your creative work and workflow! It's cheap, fast, but just imitates details. Open the SDUpscale image in a photo editor (I recommend GIMP), then open the Extras upscaled image in a layer above it. 7/24 gb until I close out of the game, which then causes it to move. It has been my go to recently, but I think it might trade some realism for sharpness. 3 times use that image for further progression Also try niam 200k upscaler,it won't give smooth like details This seems like a decent tutorial, though it doesn't seem to actually involved Stable Diffusion, it's just using the automatic1111 web ui to use an upscale and face restoration model. The process is faster, because is less complex. Take the image into inpaint mode together with all the prompts and settings and the seed. I don't like the subscription model. As mentioned above, it’s pixel perfect, that means you can literally resize the image to 32x32 using nearest neighbor and the output will be the same. What is the proper settings for upscale? I saw the big post about upscaling here, but it really changes the faces even at denoising 0. Share. It however is quite costly (could be as expensive as $1. In SD1. " Tiled VAE is a memory optimization method for the final step of the diffusion process, which is the conversion from latent space to image space using a variational autoencoder - hence, VAE. Hit gen! I've been using Foocus for most of my needs. When the picture is upscaled it introduces new details causing the tiling to not be exact any more, instead you get these very obvious "cut-line" in between tiles. 3 denoise, 25~30 steps), which gives me pretty good… My friend and I created an upscale script with the ability to use denoise 0. Of course, using latent upscale with highres fix can completely skip the conversion, so it should have some performance advantage. Reply reply More replies Here, you can see the leaves aren't pronounced and the circle is smudged. I've been using Gigapixel AI for several years on my 3D Rendered stuff as well as Conclusion. 5x 896x1152. - Is uses much less VRAM, so you will be able to use greater batch size. 51. It’s trained on upscales 32x32 to 512px It’s regular inference at 512. Plain vanilla ESRGAN 4x - surprisingly good for textures. Double check any of your upscale settings and sliders just in case. Superscale is the other general upscaler I use a lot. GIF (640x480) where it says 'drop image here'. Upscale x4 using R-ESRGAN 4x+. Hires fix uses Stable Diffusion and Stable Diffusion knows how to create images from scratch so it can add more detail. ControlNet weight at 0. The training data not only impacts the content of the image but also the composition. I can regenerate the image and use latent upscaling if that’s the best way…. The gaussian noise from the stable diffusion process gets added *after* it's converted to a latent image. There are couple good ones that 1) don't oversharpen the edges and 2) don't smudge the details. Upscale is smoothing the face which makes it look inconsistent, any advice? Upscale is smoothing the face which makes it look inconsistent, any advice to make the image more homogenous in style? I tried running it back through img2img with low denoising and the original prompt but then it jacks up the face again. So the best upscaler I have found is letsenhance (https://letsenhance. I also tried to upscale low resolution video still with the face in it, and no matter what upscaler I used, I can't say the face got of higher quality, but certainly it wasn't the man on the original picture. ControlNet Tile + ultimate SD upscale 2x Apr 5, 2023 · The first step is to get access to Stable Diffusion. Distinct-Traffic-676. Check out Remacri (gotta look around) or v4 universal (i heard is now an extension in automatic repo). Well done. So I've seen people generate a bunch of low-res images then upscale them to get a better resolution instead of spending their entire time on the…. Take that output and run it back at a higher strength. Go to controlnet, select tile_resample as my preprocessor, select the tile model. You can get just as good of results using img2img and SD upscaling, and considering you can do cleanup on your input image that way, it's ultimately going to be the better method over trying to 1 shot upscale using highres fix in img2img and hoping your initial batch setting and seed plays nice with you for this latent In the first example, notice the teddy bear, the family portrait and the ship, you can of-course add as many objects as you want, but SD upscale adds that fidelity to the image. Similar things have occurred to me. 0 - A Stable Diffusion Toolkit, a project that aims to provide enthusiasts and professionals both a suite of robust image creation tools. 5; Time: 4 x 4 seconds (16 seconds) Conclusion: More efficient workflow at 2X upscale. I haven't found a better one. Set my downsampling rate to 2 because I want more new details. 1 denoise Multi-Diffusion + Tiled VAE + ControlNet Tile will probably give you much better results than ultimate upscale. 3-0. i want to upscale an image 15 time to 11520x2160. I've got 24gb on my 3090 and sometimes it'll just hang on 23. Between them they cost more than $10. However, after creating the images I'm really lost on the best way to upscale them. The rabbit hole is pretty darn deep. I have setup Stable Diffusion on my PC, which has low-end hardware (1060 3gb). 4x Nickelback_70000G. whenever I send an image to Extras, select upscaling modes, AND enable either of the face restoration models (GFPGAN or Codeformer), I get the "ValueError: images do not match," and no image is output. It has a 1. When searching for ways to preserve skin textures, in From the Upscale Original Workflow: Dpmpp_2m, sgm_uniform, 15 steps, CFG 5. Most I had never tried before. Since moving to SDXL, my most common resolution has been 1344x1728, which is 1. Using a RTX 4090: First stage and second stage upscale combined takes about 8 minutes. None or nearest in sd upscale below 0. 7 Time: About 30-45 seconds Second upscale settings: I’m looking to upscale to 4k and probably 8k even. g. Gigapixel does a good job to the faces and skin, but nothing significant compared to open source models. Select Tab Process Image (in Vlad), Extras (in Automatic1111) Drag BARTON. jonesaid. 4. It mistakes the pixelation for blur maybe? Try bumping it up more. g Use a X2 Upscaler model. Tallied up all the images I've posted on 4chan over the last year. Thanks so much, and I'm glad that you've picked up on my intent. 2 Share. 2. 5, using one of ESRGAN models usually gives a better result in Hires Fix. 5 model (Photon and ControlNet Tile 1. My recommended workflow is pass the image through Pixelator or Use SD upscale to provide extra details. Reply. Details in the wiki. 20 for upscaling an 8K image to 16K, as I found), but the . /r/StableDiffusion is back open after the protest of Reddit killing open API Hi everyone! I was planning on doing some AI generated images to give as smalls gifts for christmas. 4x Valar. I’ll create images at 1024 size and then will want to upscale them. 5 model). This could be either because there's not enough precision to represent the picture, or because your video card does not support half type. 05 or 0. 1, and try to describe the image really well in the Steps: 150, Sampler: Euler a, CFG scale: 20, Denoising strength: 0. Does anyone have any suggestions of an upscaler that is this quality that is free or a one time purchase? Similar results of up to 16K sizes can be obtained with the new Clarity Upscaler on replicate ( philz1337x/clarity-upscaler – Run with an API on Replicate) that uses the same process under the hood, a very versatile tool in fact. 3. I've noticed that sometimes automatic1111 will stall near the end of a generation if you happen to pull up a game or something that uses up vram. But as you can see, still seams! So, I just 4x upscaled the original pic with 0. It won't add new detail to the image, but it will give you a clean upscale. It would be nice if we could see the original ones without the upscale, that way we could have a better sense on the improvements of it. WyomingCountryBoy. ControlNet Tile + ultimate SD upscale 2x (to 3072x2048). fr. . Futuristic Fusion Robots, Johnny 5 (Short Circuit), next to Tatsumaki, These two Engaged in a Heated Argument: They are face-to-face, their expressions angry and determined. 4x BS DevianceMIP_82000_G. My next most favorite, which is new Key Takeaways. The smaller size of this model makes it perfect for running on consumer PCs and laptops as well as enterprise-tier GPUs. If you're using AUTOMATIC1111's SD-UI, you can drop it into the Extras tab to upscale it. Lanczos. This was a simple 4x upscale test from 512x512 to 2048 x 2048 using several popular upscalers: BSRGAN, Lollypop, SwinIR, and Remacri. I really like using SD Ultimate Upscale for img2img but haven't found a good way to use it with ADetailer, as the tiling makes it so that ADetailer acts on each individual tile. Prompt: illustration of a It is done by resizing the picture in the latent space, so the image information must be re-generated. io/). As is to be expected, when I upscale, my people turn into plastic. It definitely alters the image a lot more, even making the flying car kind of blend in with the buildings, but it also GREATLY adds interesting, clear lettering to the signs the best approach here might be to run both ways, then combine them in a photo app to mask out some sections of the image to show the 0. Please share your tips, tricks, and workflows for using this software to create your AI art. CCSR = high fidelity, but low quality (no fine details, washed out, softens image) SUPIR = low fidelity (hallucinates too much), but very high quality (reintroduce fine details/texture) CCSR SUPIR combo is simply mind blowing as you can see in example k, l, m. I create my images using dpm++sde Karras in automatic 1111. Run one img2img pass at a low strength, clears it up a small bit. Set denoise around 0. Edit: This is a 16x upscale done iteratively. whats a good program for this /r/StableDiffusion is back open after the protest of Reddit killing open API Be the first to comment Nobody's responded to this post yet. 65 and then use SwinR-M x2 to double the resolution, so essentially a second upscale, on an already upscaled image. The first stage utilizes CCSR - 2x upscale. Enable SD upscale and crank it up to x4. For illustration/anime models you will want something smoother that I've been mainly using absolute reality with hires fix (4x ultrasharp, upscale by 2, 0. OP should have used Latent for Hires and a non-latent upscaler in SD Upscale. - It is instantaneous compared to the other upscalers. p. Just resize (latent upscale): Same as the first one, but uses latent upscaling. It makes a lot of other more complicated things super simple too. 1. /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. No upscaler. Third stage utilizes SD ULTIMATE UPSCALE - 8K size. Upscale Type: Chess. json to change the width and height slider step to 8 both in txt2img and img2img. ②i2i SD upscale-×2(1536×2048)→(3072×4096) ③i2i SD upscale-×2(3072×4096)→(6144×8192) ※ The image size of the full body image is too large to be posted, so I am downsizing and cropping it. I took several images that I rendered at 960x512, upscaled them 4x to 3840x2048, and then compared each. No need to install or download anything. What happens when you negative prompt “blur, haze”? Your prompt don't want to paint what it sees. Result will be affected by your choice relative to the amount of denoise parameter. SVD + Hires Fix Upscale (no LCM = Better Quality) + workflow. Same thing for number of iterations setting. 4x NMKD Superscale - my current favorite. Workflow: Use baseline (or generated it yourself) in img2img. Try setting the "Upcast cross attention layer to float32" option in Settings > Stable Diffusion or using the --no-half commandline argument to fix this. 3. Workflow for this image: Model: ProtovisionXL. 5 or 1. So I'm kinda new to all this but so far standard upscale tests have proven to me that topaz is way better than stable diffusion upscaled images. When I'm typically upscaling images what I notice is that the image size goes up and I'm able to zoom in further. Tbh so far multiduffusion is ridiculously slow and doesn't look visibly better than ultimate sd upscale. 5 denoise will result in a blurry/pixelated picture, using a denoise of 0. true. The depth model is from stability. Rerun those inside the inpainting tab and inpaint the whole square minus the edges. I've been upscaling images in the extras tab to 2x scale and the images don't look quite good especially around the eyes. PS: Universal Upscaler V2 - Sharper seems off color and blurrier than the others. 45 or 0. Upscale: 3072 x 4224. Where are you upscaling + what are your settings. I recommend downloading some alternate models and putting the . It generates the extra information required based on the existing image and the prompt. Mask out the extra layer, then go over your image and mask it back in over weird spots or unwanted details. This combo gave the best fidelity and quality balance. I use latent nearest exact with an upscale of 2. 000. generate your 2048x2048 image using the high-res fix, then send to extras, then upscale to 8k using any of the available options. In general it works well to take large steps at first and small steps at the end. With this I can generate images at 576 x 328, upscale the latent by 1. Step two - upscale: Change the model from the SDXL base to the refiner and process the raw picture in img2img using the Ultimate SD upscale extension with the following settings: (same prompt) Steps: 30, Sampler: DPM++ 2M Karras, CFG scale: 7, Seed: 1799556987, Size: 2304x1792, Model hash: 7440042bbd, Model: sd_xl_refiner_1. Depending on the "denoiser" setting it will alter the image more or less. Once you're satisfied, you can export the image Welcome to the unofficial ComfyUI subreddit. You do not have to think too much about a workflow when using the new tile model. then choose the what ı choosed. E. 474K subscribers in the StableDiffusion Upscale it like you did. Note the high tile width and height, padding and mask blur - all attempts to mitigate the issue, and it has definitely helped a lot from the base settings (especially changing from linear to chess actually). 1st choose extras menu then drag and drog your picture. Things: In settings-upscaling, select in Select which Real-ESRGAN models to show in the web UI. I have found if denoising is set too low on img2img this often happpens. 8192x8192 image saved as A2B. (Requires restart) the option R-ESRGAN 4x+. It seems a proccess similar to the one we can find in the EXTRAS Menu in Automatic1111 or the upscaling nodes in ComfyUI. 3s 🤯. 5 (depends on image content). ai. Stable Diffusion 3 Medium is Stability AI’s most advanced text-to-image open model yet, comprising two billion parameters. Original: 512 x 704. Sorry indeed it wasn't very obvious it's included in the second image. Use a good prompt that captures what you're after and that's all it takes. 2 and CFG scale at 14. Just testing ultimate upscaler with different pictures. I seem to be getting a lot of pans and other "basic" movement. Award. That’s because, unless you are upscaling with ControlNet Tile, you aren’t actually adding any detail. I’m struggling to find what most people are doing for this with SDXL. I then tried using SD Upscale with LCM and without, to 2x it's size: SD Upscale w/ no LCM - 60s total. Models I've tried and can confirm that this works great on: SDXL, JuggernautXL, NightvisionXL, ProtovisionXL. Enable controlnet and set preprocessor and model to tile. It seems that Upscayl only uses a upscaling model, so there is no difussion involved and the result will depend only on the upscaling model. Option 2: Use a pre-made template of Stable Diffusion WebUI on a configurable online service. The upscalers used here are: UniversalUpscalerV2-Neutral (denoted as N ) UniversalUpscalerV2-Sharp (denoted as S ProfessorTeddington. Why are you opting for the Ultimate SD Upscaler instead of upscaling with ESRGAN (after an optional initial downsize ¹), then downscaling to your preferred size, and finally using img2img? The high-res fix is for fixing the generation of high-res (>512) images. 0, VAE hash Thought I'd mention it since I haven't seen it discussed anywhere and googling "SDXL 1024x576" yields 8 (!) results. 2~0. Someone posted these settings for ControlNet Tile a few weeks back and I’ve found it works pretty well. Teaser of my upcoming pixel perfect 32x32 model. Ultimate SD upscale tile_height: initial txt2img height or one step higher Ultimate SD upscale mask_blur: 8, Upscaler chosen: UltrasharpX4 (used at X2 each time) First upscale settings:----- Denoising strength: 0. In this example, the skin of girls is better on 3rd image , because of different model used while doing img2img Ultimate SD Upscale. Definitely the best if you just wanna prompt stuff without thinking too hard. 5 (colors may change more at higher denoising), ControlNet weight: 0. which other tool can use ? or is there a batch script to latent upscale a folder via --api ??? There are extra tab for batch processing, maybe you can find something in there. Hit generate The image I now get looks exactly the same. The right upscaler will always depend on the model and style of image you are generating; Ultrasharp works well for a lot of things, but sometimes has artifacts for me with very photographic or very stylized anime models. Please keep posted images SFW. - Both 4xV3 and WDN 4xV3 are softer than x4plus. 3-1. If you don’t already have it, then you have a few options for getting it: Option 1: You can demo Stable Diffusion for free on websites such as StableDiffusion. 7, “ControlNet is more important,” ultimate SD upscale tile size set to 769x512, denoising ~0. When I started playing around with SD and other AI image generators, I really struggled to understand what any of the setting parameters actually do, since the information about them was and still is really spread out all over the place and frequently incorrect. ago. The upscaler is just used to upscale the image. A 0 won't change the image at all, and a 1 will replace it completely. But in SDXL, I find the ESRGAN models tend to oversharpen in places to give an uneven upscale. It's not bad and it's really fast. To do upscaling you need to use one of the upscaling options. Denoising around 0. Use --disable-nan-check commandline argument to System: Windows 11 64Bit, AMD Ryzen 9 3950X 16-Core Processor, 64Gb RAM, RTX3070 Ti GPU with 8Gb VRAM. Although 3x for Latent is a bit too much and not a good idea. 0; Time: 4 x 7 seconds (28 seconds) From My Workflow: LCM, sgm_uniform, 10 steps, CFG 1. My most common resolution, is 1152x1536, or double 576x768. TBH, I don't use the SD upscaler. 4x Nickelback _72000G. I've done all kinds of things trying to upscale. I've found that "Universal Upscaler v2", "Remacri", and "NMKD Siax" all work well for most things and generates good details. SD Upscale w/ LCM - 24s total. The script is the SD script from Auto1111. I tried using SD upscale (inside img2img) but the image resolution remained the same. In addition to choosing right Upscale model, it is very important to choose right model in Stable Diffusion img2img itself. I'm happy to announce the release of InvokeAI 2. The quality loss doesn't seem to be ultra noticeable, and I'm still figuring out the exact same settings for upscaling and experimenting with steps, etc etc but View community ranking In the Top 1% of largest communities on Reddit How to upscale for free Hi all, I want to upscale (with AI) an image from Stable Diffusion, 512 x 512, to a larger size. 2x upscale the base image again in the Extras tab with the same model. I don’t have the money and I use Stable Diffusion mostly for work now but there is no budget for new This workflow upscales images to 4K or 8K and upscales in 3 stages. Send that image to img2img and use the exact same prompt + the SDUpscale script + double the width & height. Running all three stages takes about 18 - 22 minutes. Add your thoughts and get the conversation going. I find using LCM with SDUpscale interesting as it causes an artistic-type effect to the resulting image, and preserves and It's too chaotic for my tastes honestly, and more a gimmick than anything. 5. OP • 2 yr. LDSR. • 1 yr. 05. Or use script SD Upscale. I'm pretty happy, so long I had a lot of errors with v-RAM when I tried to upscale pictures, and this script usually works like a charm. Use small steps for upscale, 1. 5GB of VRAM to generate a 512x768 image (and less for smaller images), and is compatible with Windows/Linux/Mac Automatic's UI has support for a lot of other upscaling models, so I tested: Real-ERSGAN 4x plus. 3 denoise strength. Just for those commenting, it doesn't matter what checkpoint you use to generate your image you can use whatever checkpoint to SD upscale it. The benefit of Hires fix (and img2img) is the option of Latent upscaling, which actually adds detail (at the cost of consistency and resolution/VRAM limitations). Loud-Preparation-212. Thanks for the help. Just follow the steps I wrote in that post and you'll be able to do the same, and probably better ! In SD 1. With yellow arrow and generate . Guessing it just generally works 👌. 5x. I have an older Mac at home and a slightly newer Mac at work. If im not mistaken latent doesn't work with pixel data, so it's not available in img2img for that reason? /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. I can generate a widescreen image with LCM + Kohya Hires (1280 x 544) in about 1. AI Model = Low Resolution. 5 or higher will create random shit in each tile, resulting in some weird fucked up chimera type thing. So, Tiled VAE isn't "infinite up scale. Ultimate SD upscale padding: 128. We are just using Ultimate SD upscales with a few control nets and tile sizes ~1024px. Im using stable diffusion web UI V1. It is suitably sized to become the next standard in text-to-image models. ComfyUI : Ultimate Upscaler - Upscale any image from Stable Diffusion, MidJourney, or photo! Using this method you can tweak as you upscale via clip - pushing in a little detail and subtle corrections as you scale up. Do SD upscale with upscaler A using 5x5 (basically 512x512 tilesize, 64 padding) [1] Send to extras, and upscale (scale 4) with upscaler B. Thanks for the workflow! Very easy to use. Second stage utilizes SUPIR - 4K size. The Karras schedule is a slight modification to the standard schedule that empirically seems to work better. Hi all, I'm looking to for the best models and workflow to upscale and apply color to old photos. Optimized for efficiency, InvokeAI needs only ~3. Title is basically the question. Twitter:@Sinori_AI----- (postscript 1) WebUI SD upscale You can change image size and add details (via denoise parameter) to skin, texture, etc Just enter new width & height and Denoising value. But for AI they are obsolete. Should add crime to negative. Also the "prompt" I'm copying has "Hires upscaler: Latent (nearest-exact)", but I don't see it. In the img2img/inpaint module, under resize mode there are 4 modes : Just resize / Crop and resize / Resize and fill ETA: if you want to use automatic, there is no longer a hard restriction for multiples of 64 anymore. If you're in the mood to experiment, you can drop it in the img2img tab, keep the Denoising strength really low, like 0. Ultimate SD upscale is great for upscaling but not when it's a tiling image/texture, in my experience. If you want to check the whole picture, please check my Twitter. Then you'll want to get the one you like best and Depends on the image, usually. Topaz Labs Gigapixel settings: Scale = 6x. that can add details while preserving the original look of the image and really effective when using vae like mse-840000-ema. Why are you not using tiled vae along with tiled diffusion If you want to add objects use break word and use If you want more details my suggestion is initially dont directly upscale to 2times instead do 1. Hope that clears up the confusion. It can be useful for two reasons : - It can add more details than a normal upscaler. scale it to whatever size I want. If you want the sharpest image, Remacri wins easily, but it's almost too sharp. pth file for each model in the "stable-diffusion-webui\ESRGAN" folder, then restarting the app. The new hard restriction is multiples of 8. resrgan is alright, but removes texture and makes hair look like clothes. Full prompt and seed: !dream "white marble interior photograph, architecture carved, shiny, brutalist, smooth, expansive, by louis kahn and moshe safdie " -H 704 -n 9 -i -S 3575419545. I believe it should work with 8GB vram provided your SDXL Model and Upscale model are not super huge E. Then afterwards pull out the pieces in something like Photopea (I used Clip Studio Paint, but any image editor you're comfortable with works) in chunks of either 512, 768, or 1024 panels. Enter NVIDIA RTX Remix, a free modding platform built on NVIDIA Omniverse that enables modders to quickly create and share #RTXON mods for classic games, each with enhanced materials, full ray tracing, NVIDIA DLSS 3, and NVIDIA Reflex. Among 2452 images, My average resolution is 1438x1729. 4x NMKD Siax - also very good. Step 1: Initial upscale. Which one is better will depend on Generate a 512xwhatever image which I like. Also, I liked the better result, using a 1. 3 - 0. you have freely upscaled your picture. Yes, you can use whatever model you want when running img2img; the trick is how much you denoise. s. Buying anything new is not in the cards for a couple of years. My goto upscale method for Hires Fix in SDXL is good old Lanczos which gives me a clean and even upscale. Also, you can downsample images in photoshop too. Then upscale the final product. Maybe try to be more descriptive about the picture. Find a chainner on github, it lets you use various upscale py models. Great for graphic design and photography. 4, cfg ~4-4. Zealousideal_Royal14. It regenerates the input image with a larger resolution. mv aa px fv sq un fm xx dg zz