minor summar updates
Browse files
README.md
CHANGED
@@ -12,17 +12,21 @@ language:
|
|
12 |
---
|
13 |
# CommonCanvas-S-NC
|
14 |
|
|
|
|
|
15 |
## Summary
|
16 |
CommonCanvas is a family of latent diffusion models capable of generating images from a given text prompt.
|
17 |
Different models in the family are different sizes, and trained on different subsets of the CommonCatalog Dataset (See Data Card), a large dataset of Creative Commons licensed images with synthetic captions produced using a pre-trained BLIP-2 captioning model.
|
18 |
CommonCanvas-S-NC is the small (S) model based off the Stable Diffusion 2 architecture, and trained on the non-commercial (NC) subset of CommonCatalog.
|
19 |
|
|
|
|
|
|
|
|
|
|
|
20 |
**Input:** CommonCatalog Text Captions
|
21 |
**Output:** CommonCatalog Images
|
22 |
**Architecture:** Stable Diffusion 2
|
23 |
-
**Version Number:** 0.1
|
24 |
-
|
25 |
-
The goal of this purpose is to produce a model that is competitive with Stable Diffusion 2, but to do so using an easily accessible dataset of known provenance. Doing so makes replicating the model significantly easier and provides proper attribution to all the creative commons work used to train the model. The exact training recipe of the model can be found in the paper hosted at this link. https://arxiv.org/abs/2310.16825
|
26 |
|
27 |
## Performance Limitations
|
28 |
|
|
|
12 |
---
|
13 |
# CommonCanvas-S-NC
|
14 |
|
15 |
+
**Version Number:** 0.1
|
16 |
+
|
17 |
## Summary
|
18 |
CommonCanvas is a family of latent diffusion models capable of generating images from a given text prompt.
|
19 |
Different models in the family are different sizes, and trained on different subsets of the CommonCatalog Dataset (See Data Card), a large dataset of Creative Commons licensed images with synthetic captions produced using a pre-trained BLIP-2 captioning model.
|
20 |
CommonCanvas-S-NC is the small (S) model based off the Stable Diffusion 2 architecture, and trained on the non-commercial (NC) subset of CommonCatalog.
|
21 |
|
22 |
+
The goal of this purpose is to produce a high-quality text-to-image model, but to do so using an easily accessible dataset of known provenance.
|
23 |
+
The exact training recipe of the model can be found in the paper hosted at this link. https://arxiv.org/abs/2310.16825
|
24 |
+
|
25 |
+
|
26 |
+
### Training Overview
|
27 |
**Input:** CommonCatalog Text Captions
|
28 |
**Output:** CommonCatalog Images
|
29 |
**Architecture:** Stable Diffusion 2
|
|
|
|
|
|
|
30 |
|
31 |
## Performance Limitations
|
32 |
|