Qwen Image Lora not capturing enough details #523
Replies: 6 comments 3 replies
-
Beta Was this translation helpful? Give feedback.
-
|
It depends heavily on the dataset, so it's hard to say for sure. How many training images are there and are they consistent? Also, what are the captions like and are the captions for generation similar to them? Also, dim=128 seems too high, and personally, I always set alpha to 1 to simplify learning rate estimation. And you might want to increase the number of steps a bit. I think you could start with simple settings, such as lr_scheduler=constant,dim=16, alpha=1, learning_rate=1e-4~1e-3. Also, try setting timestep_sampling to shift and discrete_flow_shift to 2.2, or timestep_sampling to uniform. |
Beta Was this translation helpful? Give feedback.
-
|
I found that cosine with 3 restarts worked well for my use-case before but let me try simplifying my settings to see how that works. Also, my understanding of alpha was that it controlled the "strength" of the training. So, if I set alpha to 1, would that mean that it will require many more steps to train? Please forgive me if I'm wrong but, since I want to capture the "details" of the image, would decreasing rank be counter productive? To further clarify: I have 3 images taken of the furniture from different angles in a studio setting. The captions are like: The generation is like: I found that qwen lora training is able to capture the shape well. But it misses out on small details that when I trained on flux it won't miss. (similar to what Furkan says) |
Beta Was this translation helpful? Give feedback.
-
|
here few my top quality so far still doing R&D
|
Beta Was this translation helpful? Give feedback.
-
|
I gave up, whatever I tried skin just comes up smooth, cannot get the detail into it. |
Beta Was this translation helpful? Give feedback.
-
|
Be sure to test your resulting loras at something like 50 steps - that helps a lot with details. I've found timestep_sampling set to sigmoid is good for most things - you can go uniform for the last bits of training to really bake in details but I wouldn't recommend it the whole time. Train at 640x640 and 1328x1328 buckets. |
Beta Was this translation helpful? Give feedback.







Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I'm trying to create Loras furniture. For example, a cabinet.
I am able to generally get the lora to capture the cabinet's structure well. But, the details of the cabinet is sometimes wrong. For example, the grain of the wood could be wrong.
My hypothesis is that I need to train more on the later timesteps (focused more on details), but I can't seem to get it working. I even tried the qinglong_qwen timestep_sampling option that supposed to train more on details but that didn't work.
Would appreciate any tips
Beta Was this translation helpful? Give feedback.
All reactions