Commit
·
927d798
1
Parent(s):
9fdbbd4
Update README.md
Browse files
README.md
CHANGED
@@ -15,14 +15,12 @@ stable-diffusion-webui\embeddings and add "boring_e621_v4" to your negative prom
|
|
15 |
|
16 |
## Model Description
|
17 |
|
18 |
-
|
19 |
The motivation for boring_e621 is that negative embeddings like [Bad Prompt](https://huggingface.co/datasets/Nerfgun3/bad_prompt),
|
20 |
whose training is described [here](https://www.reddit.com/r/StableDiffusion/comments/yy2i5a/i_created_a_negative_embedding_textual_inversion/)
|
21 |
depend on manually curated lists of tags describing features people do not want their images to have, such as "deformed hands". Some problems with this approach are:
|
22 |
* Manually compiled lists will inevitably be incomplete.
|
23 |
* Models might not always understand the tags well due to a dearth of training images labeled with these tags.
|
24 |
* It can only capture named concepts. If there exist unnamed yet visually unappealing concepts that just make an image look wrong, but for reasons that cannot be succinctly explained, they will not be captured by a list of tags.
|
25 |
-
<br>
|
26 |
|
27 |
To address these problems, boring_e621 employs textual inversion on a set of images automatically extracted from the art site
|
28 |
e621.net, a rich resource of millions of hand-labeled artworks, each of which is both human-labeled topically and rated
|
@@ -45,7 +43,7 @@ To qualitatively evaluate how well boring_e621 has learned to improve image qual
|
|
45 |
![boring_e621 and boring_e621_v4 Performance on Simple Prompts](tmpoqs1d_vv.png)
|
46 |
|
47 |
As we can see, putting these embeddings in the negative prompt yields a more delicious burger, a more vibrant and detailed landscape, a prettier pharoah, and a more 3-d-looking aquarium.
|
48 |
-
|
49 |
|
50 |
## Other Models
|
51 |
|
|
|
15 |
|
16 |
## Model Description
|
17 |
|
|
|
18 |
The motivation for boring_e621 is that negative embeddings like [Bad Prompt](https://huggingface.co/datasets/Nerfgun3/bad_prompt),
|
19 |
whose training is described [here](https://www.reddit.com/r/StableDiffusion/comments/yy2i5a/i_created_a_negative_embedding_textual_inversion/)
|
20 |
depend on manually curated lists of tags describing features people do not want their images to have, such as "deformed hands". Some problems with this approach are:
|
21 |
* Manually compiled lists will inevitably be incomplete.
|
22 |
* Models might not always understand the tags well due to a dearth of training images labeled with these tags.
|
23 |
* It can only capture named concepts. If there exist unnamed yet visually unappealing concepts that just make an image look wrong, but for reasons that cannot be succinctly explained, they will not be captured by a list of tags.
|
|
|
24 |
|
25 |
To address these problems, boring_e621 employs textual inversion on a set of images automatically extracted from the art site
|
26 |
e621.net, a rich resource of millions of hand-labeled artworks, each of which is both human-labeled topically and rated
|
|
|
43 |
![boring_e621 and boring_e621_v4 Performance on Simple Prompts](tmpoqs1d_vv.png)
|
44 |
|
45 |
As we can see, putting these embeddings in the negative prompt yields a more delicious burger, a more vibrant and detailed landscape, a prettier pharoah, and a more 3-d-looking aquarium.
|
46 |
+
<br>
|
47 |
|
48 |
## Other Models
|
49 |
|