StableDiffusion

98 readers
1 users here now

/r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and...

founded 1 year ago
MODERATORS
801
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/Familiar-Art-6233 on 2024-08-27 00:51:30+00:00.

802
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/mercantigo on 2024-08-27 00:23:21+00:00.

803
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/CeFurkan on 2024-08-26 23:34:25+00:00.

804
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/IAintNoExpertBut on 2024-08-26 23:13:11+00:00.

805
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/Nyao on 2024-08-26 22:19:06+00:00.

806
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/tabula_rasa22 on 2024-08-26 21:35:09+00:00.

807
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/More_Bid_2197 on 2024-08-26 18:21:47+00:00.


Be careful with the folder and the name of the models

I'm not sure, but I think that Clip L in the Black labs directory does not work for training (the correct model is 234.7 MB)

Error 1 - T5XXL - you need to download the Fp16 version. It does not work with FP8

Error 2 - location of the models

Place all the models in the "models" folder. Do not place the models in another folder. And use the following path in Kohya.

./models/flux1-dev.safetensors

When you add this path in "pretrained model path" and press the spin arrow button, Kohya will display the Flux options. IMPORTANT

Then, you have to add the path of the other 3 models further down (much further down)

./models/ae.safetensors, do the same for ./models/clip_l.safetensors, and ./models/t5xxl_fp16.safetensors

DO NOT CHANGE THE NAMES

attention - it's called AE in Flux

2 - second most common mistake, image folder

Folder with images (example - photos) and inside this folder there must be another folder, for example, with the name 1_ohwx man. And the images are there The number indicates the number of repetitions

3 - Don't forget to click on "folders". And choose a folder for outputs. There is a folder in the kohya directory with this name

4 - Be careful with Adamw 8 bits. It didn't work for me. You need to install another package bitsandbytes (I tried but couldn't), use Adam normal, Adafactor, prodigy

5 - Enable the options

Cache text encoder outputs to speed up inference

Cache text encoder outputs to disk to speed up inference

I can't explain why, but an error appeared when it was disabled (I'm not sure if you need both, at least the first one)

7 - In model predict use the raw option

And in Timestep Sampling - Sigmoid

Without this it didn't work and I can't explain why

8- With 24 GB of Vram I only got the resolution 512 X 512 with the option fp8 for base model (there is another option that I checked and it allowed me to train in Bf16 but it became extremely slow, it would take more than 9 hours, so for me it's not a good idea)

9 - DO NOT click on Parameters PRESETS. When selecting kohya's default settings for Flux, a bug was generated saying that the model was not located (something like model.safetensors/r) the r doesn't make sense. When clicking on this the BUG persists even if you remove it.

10 - I trained with GPU online. Maybe windows is different

11 - DO NOT open runpod directly on port 7860. The GUI will appear, but the GUI alone does not show the training.

You need to select the image with kohya and enter the jupyter notebook (usually port 8888). And after that, go to the kohya directory and run the command

cd kohya_ss

./setup-runpod.sh (to run the script, it may not be necessary)

./gui.sh --share --headless (to launch the GUI, after this command a gradio link will appear at the end to access the gui)

Keep the jupyter notebook open to see if it is training, errors, etc.

808
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/FlavoredQuark on 2024-08-26 16:59:32+00:00.

809
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/the_bollo on 2024-08-26 18:39:04+00:00.

810
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/OkSpot3819 on 2024-08-26 17:49:46+00:00.

811
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/WINDOWS91 on 2024-08-26 16:42:35+00:00.

812
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/idunno63 on 2024-08-26 15:09:48+00:00.

813
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/haofanw on 2024-08-26 14:44:12+00:00.


The InstantX Team and Shakker Labs jointly release two new ControlNets.

814
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/AggravatingStable490 on 2024-08-26 13:48:03+00:00.

815
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/LatentDimension on 2024-08-26 10:57:41+00:00.

816
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/rolux on 2024-08-26 10:57:01+00:00.

817
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/Pyros-SD-Models on 2024-08-26 14:20:21+00:00.


I promised you a high quality lewd FLUX fine-tune, but, my apologies, that thing's still in the cooker because every single day, I discover something new with flux that absolutely blows my mind, and every other single day I break my model and have to start all over :D

In the meantime I've written down some of these mind-blowers, and I hope others can learn from them, whether for their own fine-tunes or to figure out even crazier things you can do.

If there’s one thing I’ve learned so far with FLUX, it's this: We’re still a good way off from fully understanding it and what it actually means in terms of creating stuff with it, and we will have sooooo much fun with it in the future :)

Any questions? Feel free to ask or join my discord where we try to figure out how we can use the things we figured out for the most deranged shit possible. jk, we are actually pretty SFW :)

818
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/luke__uk on 2024-08-26 11:40:00+00:00.

819
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/Kinfolk0117 on 2024-08-26 08:27:42+00:00.

820
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/julimoooli on 2024-08-26 11:07:48+00:00.


I've noticed something interesting about the Flux model—one of its standout features for me is the way it produces unique faces and anatomies, aside from the occasional cleft chin. In the past, it was easy to identify AI-generated images at a glance, even before scrutinizing the hands or other imperfections, just by recognizing the distinct "1girl" face. Fortunately, with Flux, this issue seems to be partly resolved.

 However, while browsing Civitai today, I observed that many NSFW LoRas are generating faces and body parts that look almost identical to those produced by Pony Realism and SDXL models. And here's the kicker—I downloaded a dataset from one of these LoRas, and the training images were actually generated by Pony. Now, don't get me wrong—I have nothing against Pony. I've had a lot of fun using it, and it's brilliantly fine-tuned for its purpose.

 But as an average user experimenting and having fun with generative AI, I can't help but wonder if we're heading towards a situation where these LoRas get merged into Flux models, and then other models get merged based on those, and so on. You see where I'm going with this, right? It's the same cycle we've seen with many SD 1.5, SDXL, and Pony merges.

 Again, this is just my observation, and since I'm not a professional in this area, I'd love to hear your thoughts and predictions. What do you think?

821
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/MGS023 on 2024-08-26 10:55:46+00:00.

822
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/Kinda-Brazy on 2024-08-26 10:23:02+00:00.

823
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/rolux on 2024-08-26 09:24:50+00:00.

824
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/Outrageous-Text-9233 on 2024-08-26 05:41:20+00:00.


We are happy to release a sdxl lineart controlnet.

It's trained on carefully selected proprietary real-world images, with 100k training steps (with batchsize 4).

This controlnet was used by promeai.pro to enpower thousands of designers and marketers to simplify their workflow and increase the productivities.

Check huggingface repo to start easily.

here are some examples.

825
 
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/love1008 on 2024-08-26 05:23:52+00:00.

view more: ‹ prev next ›