“Better quality” is an interesting concept. Increasing steps, depending in the sampler, changes the image. The seed mode usually changes image with changes in size.
So, what exactly do you mean with “better quality”?
“Better quality” is an interesting concept. Increasing steps, depending in the sampler, changes the image. The seed mode usually changes image with changes in size.
So, what exactly do you mean with “better quality”?
This isn’t what you asked specifically, but it’s related enough… have a look into https://apps.apple.com/it/app/draw-things-ai-generation/id6444050820?l=en-GB as it’s free, ad free, free from tracking and really well optimized. With that I can run Schnell on my iPhone 13 Pro!
I enjoy this 1.5 Lora https://civitai.com/models/165876/2d-pixel-toolkit-2d it’s pretty neat!
The floor is a carpet, and the shoes is harder to tell but might be a similar situation? Velvet maybe?
I’m guessing there’s a mix. The smallest version is 700 million, possibly the one used to generate the time data reported, but the largest (or not?) still runs with 8gb. If I remember correctly SD3 is supposed to have multiple versions, starting from 800 millions and going up, so this is going to be interesting.
Cool, looks simple enough. Can’t test it on my phone, but for things with the A12 and up (although ram can and will be a problem if less than 6gb) there’s https://apps.apple.com/it/app/draw-things-ai-generation/id6444050820?l=en-GB
Can I offer what I believe is a better option? 1.5 LCM models. 5 steps for a good image, and they’re 5 steps at 1.5 speeds :)
I like this but obviously you can find other LCM models.
How old of a Xeon? Because it won’t be a fast result, but maybe you are fine with it. Back when I tried this, SD 1.5 could do 20 steps at 512x512 on my Ryzen 5 3600 in roughly 7 minutes…
Unless they aim for a specialized model? I don’t have insight on the matter, just a guess.
Is it? The authors all have names that (in my ignorance!) all sound Russian, and Kandinsky was a Russian painter…
Yup, same reason why you can ask for a fox using a crocodile as a mech and get a good result. The model has the concept of all things requested and mixes them (with varying success).
Much simpler way, even if not as cool: try https://apps.apple.com/app/id6444050820 that is free and great. First download a model, if you already have one you like just use the Civitai.com link, otherwise there’s a selection of them and I’d try… Realistic Visions? The 8 bit one saves you a little space. Then use the photo icon in the bottom to import a photo, you want something with the face clearly visible. Use the eraser to get rid of everything other than the face. Now try prompting for an astronaut! Should work already but you can try adding Controls: Tile, Inpainting and Pose specifically.
Performance wise you can go at the bottom of the left side, click on the cpu looking icon and select (if it’s not selected already) Use CoreML: Yes and for the CoreML compute units go with All.
My M2 Mini takes a dozen seconds for a normal generation like that. With the Control modules and Inpainting is going to be a little slower obviously but hey, it’s a fast app.
Not going to lie, I almost gave up a few times. But I can also be stubborn… anyway since this is apparently the first confirmation it works, it’s probably be helpful if I mention that it’s a 12gb 3060. :)
My request is dumb, the UI is glitching a little but hot damn 12 iterations per second! Impressive.
… I’ll check later, but I do remember grabbing the “right one” as I had version 12, so this might very well be it.
Now knowing where to look, I did some fixing by myself! Main issue is that I had CUDA 10 and 12, no 11. Then after going insane about that tiny difference… I landed on something I lack the knowledge to decipher: “PyInstallerImportError: Failed to load dynlib/dll ‘C:\Program Files\NVIDIA GPU Computing Toolkit\TensorRT-8.6.1.6\lib\nvinfer_plugin.dll’. Most likely this dynlib/dll was not found when the application was frozen.”
All I can say is that the file is there.
So, feedback. To begin with, it works! That’s a massive improvement and allowed me to actually try it. Civitai.com downloading works quite nicely and… the generation is kinda slow. Slower than my iPhone 13 pro with Draw Things, a minute give or take 10 seconds. Poor phone crunches the same model in 30 something seconds.
Don’t get me wrong, I appreciate it works to begin with, it’s also easy to setup, but there’s a fair amount of performance left on the table. Now depending on how much work there’s to do it might make sense to chase further performance, but that’s something only you can decide :D
3060 here, it might be the vram. SDXL eats a lot of it (and if you had say the vae in the wrong spot it would output very wrong images) so it might be that either 8gb aren’t enough, or maybe they aren’t enough with the resolution of your screen plus whatever you are running, like the browser.
Or, OR: the checkpoint is corrupted. I had that happen a couple of times in the past and the whole huge error with loading of another model was what happened.
Really glad that could help! Since I’ve got your attention, I couldn’t get TensorRT to work on Windows. At least 50% chance I didn’t install it properly, BUT at the same time your gui was showing my 1650 instead of the 3060. After looking for some setting about Cuda devices and finding none I gave up. Generation times and usage pointed clearly at a normal 3060 task, even if the gui had the temperature for the 1650.
But anyway! One thing I’d like to ask is (now that there’s a viable way to use it on my Mini) an option to allow other computers to access it, and better yet the API like in Automatic1111. Like that I could do some kind of LLM on the 3060 (I like Pygmalion 6b) and stable diffusion on the Mac.
All that aside, thanks for making a viable alternative to Draw Things. As much as I like it and the interface, choice is always good… and yours has the potential to be usable in remote :D
I have a few examples that I hope retain their metadata.
Seed mode is… basically, I stopped using Automatic1111 a long time ago and kinda lost track of what goes on there but in the app I use (Draw Things) there’s a seed mode called Scale Alike. Could be exclusive, could be the standard everywhere for what I know. It does what it says, changing resolution will keep things looking close enough.
Edit: obviously at some point they had to lose the bloody metadata….