Upload 11 files
Browse files- README.md +40 -2
- config.json +21 -0
- merges.txt +0 -0
- special_tokens_map.json +30 -0
- text_encoder.bin +3 -0
- text_encoder.xml +0 -0
- tokenizer_config.json +30 -0
- vae_decoder.bin +3 -0
- vae_decoder.xml +0 -0
- vae_encoder.bin +3 -0
- vae_encoder.xml +0 -0
README.md
CHANGED
|
@@ -1,6 +1,44 @@
|
|
| 1 |
---
|
| 2 |
-
license:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
# Uses
|
| 5 |
|
| 6 |
## Direct Use
|
|
@@ -58,4 +96,4 @@ ability of the model to generate content with non-English prompts is significant
|
|
| 58 |
|
| 59 |
|
| 60 |
### Intel’s Human Rights Disclaimer:
|
| 61 |
-
Intel is committed to respecting human rights and avoiding complicity in human rights abuses. See Intel's Global Human Rights Principles. Intel's products and software are intended only to be used in applications that do not cause or contribute to a violation of an internationally recognized human right.
|
|
|
|
| 1 |
---
|
| 2 |
+
license: creativeml-openrail-m
|
| 3 |
+
extra_gated_prompt: |-
|
| 4 |
+
This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
|
| 5 |
+
The CreativeML OpenRAIL License specifies:
|
| 6 |
+
1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
|
| 7 |
+
2. Intel claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
|
| 8 |
+
3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
|
| 9 |
+
Please read the full license carefully here: https://huggingface.co/spaces/CompVis/stable-diffusion-license
|
| 10 |
+
|
| 11 |
+
extra_gated_heading: Please read the LICENSE to access this model
|
| 12 |
---
|
| 13 |
+
# SD v1-5 square Model Card
|
| 14 |
+
The original source of this model is : [runwayml/stable-diffusion-v1-5](https://huggingface.co/runwayml/stable-diffusion-v1-5).
|
| 15 |
+
This model is just optimized and converted to Intermediate Representation (IR) using OpenVino's Model Optimizer and POT tool to run on Intel's Hardware - CPU, GPU, NPU.
|
| 16 |
+
|
| 17 |
+
We have FP16 and INT8 versions of the model. Please note currently only unet model is quantized to int8.
|
| 18 |
+
|
| 19 |
+
Intended to be used with:
|
| 20 |
+
- GIMP plugin [openvino-ai-plugins-gimp](https://github.com/intel/openvino-ai-plugins-gimp.git)
|
| 21 |
+
- Blender Addon [dream-textures-openvino](https://github.com/intel/dream-textures-openvino)
|
| 22 |
+
|
| 23 |
+
## Original Model Details
|
| 24 |
+
- **Original Developed by:** Robin Rombach, Patrick Esser
|
| 25 |
+
- **Model type:** Diffusion-based text-to-image generation model
|
| 26 |
+
- **Language(s):** English
|
| 27 |
+
- **License:** [The CreativeML OpenRAIL M license](https://huggingface.co/spaces/CompVis/stable-diffusion-license) is an [Open RAIL M license](https://www.licenses.ai/blog/2022/8/18/naming-convention-of-responsible-ai-licenses), adapted from the work that [BigScience](https://bigscience.huggingface.co/) and [the RAIL Initiative](https://www.licenses.ai/) are jointly carrying in the area of responsible AI licensing. See also [the article about the BLOOM Open RAIL license](https://bigscience.huggingface.co/blog/the-bigscience-rail-license) on which our license is based.
|
| 28 |
+
- **Model Description:** This is a model that can be used to generate and modify images based on text prompts. It is a [Latent Diffusion Model](https://arxiv.org/abs/2112.10752) that uses a fixed, pretrained text encoder ([CLIP ViT-L/14](https://arxiv.org/abs/2103.00020)) as suggested in the [Imagen paper](https://arxiv.org/abs/2205.11487).
|
| 29 |
+
- **Resources for more information:** [GitHub Repository](https://github.com/CompVis/stable-diffusion), [Paper](https://arxiv.org/abs/2112.10752).
|
| 30 |
+
- **Cite as:**
|
| 31 |
+
|
| 32 |
+
@InProceedings{Rombach_2022_CVPR,
|
| 33 |
+
author = {Rombach, Robin and Blattmann, Andreas and Lorenz, Dominik and Esser, Patrick and Ommer, Bj\"orn},
|
| 34 |
+
title = {High-Resolution Image Synthesis With Latent Diffusion Models},
|
| 35 |
+
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
|
| 36 |
+
month = {June},
|
| 37 |
+
year = {2022},
|
| 38 |
+
pages = {10684-10695}
|
| 39 |
+
}
|
| 40 |
+
|
| 41 |
+
|
| 42 |
# Uses
|
| 43 |
|
| 44 |
## Direct Use
|
|
|
|
| 96 |
|
| 97 |
|
| 98 |
### Intel’s Human Rights Disclaimer:
|
| 99 |
+
Intel is committed to respecting human rights and avoiding complicity in human rights abuses. See Intel's Global Human Rights Principles. Intel's products and software are intended only to be used in applications that do not cause or contribute to a violation of an internationally recognized human right.
|
config.json
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"power modes supported": "yes",
|
| 3 |
+
"best performance": [
|
| 4 |
+
"GPU",
|
| 5 |
+
"GPU",
|
| 6 |
+
"GPU",
|
| 7 |
+
"GPU"
|
| 8 |
+
],
|
| 9 |
+
"balanced": [
|
| 10 |
+
"NPU",
|
| 11 |
+
"NPU",
|
| 12 |
+
"GPU",
|
| 13 |
+
"GPU"
|
| 14 |
+
],
|
| 15 |
+
"best power efficiency": [
|
| 16 |
+
"NPU",
|
| 17 |
+
"NPU",
|
| 18 |
+
"NPU",
|
| 19 |
+
"GPU"
|
| 20 |
+
]
|
| 21 |
+
}
|
merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<|startoftext|>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": true,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "<|endoftext|>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "<|endoftext|>",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"unk_token": {
|
| 24 |
+
"content": "<|endoftext|>",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
}
|
| 30 |
+
}
|
text_encoder.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f245795911e496f451a01e1a5de70578753c985ed76ffa302966d2e716d2b8d0
|
| 3 |
+
size 246133100
|
text_encoder.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"added_tokens_decoder": {
|
| 4 |
+
"49406": {
|
| 5 |
+
"content": "<|startoftext|>",
|
| 6 |
+
"lstrip": false,
|
| 7 |
+
"normalized": true,
|
| 8 |
+
"rstrip": false,
|
| 9 |
+
"single_word": false,
|
| 10 |
+
"special": true
|
| 11 |
+
},
|
| 12 |
+
"49407": {
|
| 13 |
+
"content": "<|endoftext|>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": false,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false,
|
| 18 |
+
"special": true
|
| 19 |
+
}
|
| 20 |
+
},
|
| 21 |
+
"bos_token": "<|startoftext|>",
|
| 22 |
+
"clean_up_tokenization_spaces": true,
|
| 23 |
+
"do_lower_case": true,
|
| 24 |
+
"eos_token": "<|endoftext|>",
|
| 25 |
+
"errors": "replace",
|
| 26 |
+
"model_max_length": 77,
|
| 27 |
+
"pad_token": "<|endoftext|>",
|
| 28 |
+
"tokenizer_class": "CLIPTokenizer",
|
| 29 |
+
"unk_token": "<|endoftext|>"
|
| 30 |
+
}
|
vae_decoder.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:907ab331d35ce61c6732a0802372d1f809e1dea27184c7297277d743cf6ca206
|
| 3 |
+
size 98980680
|
vae_decoder.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
vae_encoder.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8170c14a8904c820a2fe3745e8a292118cd15e4805c448c18959524fa697d257
|
| 3 |
+
size 68327548
|
vae_encoder.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|