Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,38 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
tags:
|
4 |
+
- diffusion
|
5 |
+
- style_similarity
|
6 |
+
- CSD
|
7 |
+
language:
|
8 |
+
- en
|
9 |
+
pipeline_tag: text2text-generation
|
10 |
+
license: cc-by-4.0
|
11 |
+
---
|
12 |
+
|
13 |
+
# Quick Links
|
14 |
+
|
15 |
+
- **GitHub Repository**: https://github.com/learn2phoenix/CSD
|
16 |
+
- **arXiv**: https://arxiv.org/abs/2404.01292
|
17 |
+
|
18 |
+
# Description
|
19 |
+
|
20 |
+
We present a framework for understanding and extracting style descriptors from images. Our framework comprises a new dataset curated using the insight that style is a subjective property
|
21 |
+
of an image that captures complex yet meaningful interactions of factors including but not limited to colors, textures, shapes, etc.We also propose a method to extract
|
22 |
+
style descriptors that can be used to attribute style of a generated image to the images used in the training dataset of a text-to-image mode
|
23 |
+
|
24 |
+
# Technical Specification
|
25 |
+
The checkpoint is for ViT-Large model
|
26 |
+
|
27 |
+
# Cite our work
|
28 |
+
|
29 |
+
If you find our model, codebase or dataset beneficial, please consider citing our work:
|
30 |
+
|
31 |
+
```bibtex
|
32 |
+
@article{somepalli2024measuring,
|
33 |
+
title={Measuring Style Similarity in Diffusion Models},
|
34 |
+
author={Somepalli, Gowthami and Gupta, Anubhav and Gupta, Kamal and Palta, Shramay and Goldblum, Micah and Geiping, Jonas and Shrivastava, Abhinav and Goldstein, Tom},
|
35 |
+
journal={arXiv preprint arXiv:2404.01292},
|
36 |
+
year={2024}
|
37 |
+
}
|
38 |
+
```
|