[Question] How is prompt attention different compare to A1111's? #3041
-
Recently I switched to SD.Next from A1111's webui because I want to try new stuffs and I heard SD.Next is faster and yes it's sooooo much faster. But I couldn't help but notice that prompting is a bit different in SD.Next, specifically applying more attention to the prompt e.g. People are saying the Diffusers backend is different but can anyone explain how it's different and what're the best practices for applying attention in SD.Next? I mainly use SDXL models. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
rule-of-a-thumb is that attention sum should be neutral. if you're bringing attention to something and not reducing attention from something else your prompt becomes unbalanced and can result in what appears as burnt image.
you can get away with slight unbalancing, so even worse if you apply it to negative prompt. contrary common belief, negative prompts don't work by preventing something to start with - they ADD it first and then steer away from it. but in general, with modern prompt parser and modern models, there is faaar less need for attention than it used to be. |
Beta Was this translation helpful? Give feedback.
rule-of-a-thumb is that attention sum should be neutral. if you're bringing attention to something and not reducing attention from something else your prompt becomes unbalanced and can result in what appears as burnt image.
a (cat:1.5) and a (dog:0.5)
would be an example of balanced prompt.you can get away with slight unbalancing, so
a (cat:1.1)
is still fine, buta (cat:1.5)
means majority of prompt is under very strong attention and that doesnt look nicely.even worse if you apply it to negative prompt.
contrary common belief, negative prompts don't work by preventing something to start with - they ADD it first and then steer away from it.
so if you add
a (cat:1.5)
, you just added a gh…