You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+24-15
Original file line number
Diff line number
Diff line change
@@ -5,7 +5,7 @@
5
5
## News
6
6
7
7
**July 26, 2023**
8
-
- We are releasing two new open models with a permissive [`CreativeML Open RAIL++-M` license](model_licenses/LICENSE-SDXL1.0):
8
+
- We are releasing two new open models with a permissive [`CreativeML Open RAIL++-M` license](model_licenses/LICENSE-SDXL1.0) (see [Inference](#inference) for file hashes):
9
9
-[SDXL-base-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0): An improved version over `SDXL-base-0.9`.
10
10
-[SDXL-refiner-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0): An improved version over `SDXL-refiner-0.9`.
11
11
@@ -22,11 +22,11 @@
22
22
-`SDXL-base-0.9`: The base model was trained on a variety of aspect ratios on images with resolution 1024^2. The base model uses [OpenCLIP-ViT/G](https://github.com/mlfoundations/open_clip) and [CLIP-ViT/L](https://github.com/openai/CLIP/tree/main) for text encoding whereas the refiner model only uses the OpenCLIP model.
23
23
-`SDXL-refiner-0.9`: The refiner has been trained to denoise small noise levels of high quality data and as such is not expected to work as a text-to-image model; instead, it should only be used as an image-to-image model.
24
24
25
-
If you would like to access these models for your research, please apply using one of the following links:
26
-
[SDXL-0.9-Base model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-0.9-Refiner](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
27
-
This means that you can apply for any of the two links - and if you are granted - you can access both.
25
+
If you would like to access these models for your research, please apply using one of the following links:
26
+
[SDXL-0.9-Base model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-0.9-Refiner](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
27
+
This means that you can apply for any of the two links - and if you are granted - you can access both.
28
28
Please log in to your Hugging Face Account with your organization email to request access.
29
-
**We plan to do a full release soon (July).**
29
+
**We plan to do a full release soon (July).**
30
30
31
31
## The codebase
32
32
@@ -58,12 +58,12 @@ cd generative-models
58
58
59
59
#### 2. Setting up the virtualenv
60
60
61
-
This is assuming you have navigated to the `generative-models` root after cloning it.
61
+
This is assuming you have navigated to the `generative-models` root after cloning it.
62
62
63
63
**NOTE:** This is tested under `python3.8` and `python3.10`. For other python versions, you might encounter version conflicts.
64
64
65
65
66
-
**PyTorch 1.13**
66
+
**PyTorch 1.13**
67
67
68
68
```shell
69
69
# install required packages from pypi
@@ -72,7 +72,7 @@ source .pt13/bin/activate
72
72
pip3 install -r requirements/pt13.txt
73
73
```
74
74
75
-
**PyTorch 2.0**
75
+
**PyTorch 2.0**
76
76
77
77
78
78
```shell
@@ -112,13 +112,22 @@ You will find the built package in `dist/`. You can install the wheel with `pip
112
112
Note that the package does **not** currently specify dependencies; you will need to install the required packages,
113
113
depending on your use case and PyTorch version, manually.
114
114
115
-
## Inference:
115
+
## Inference
116
116
117
-
We provide a [streamlit](https://streamlit.io/) demo for text-to-image and image-to-image sampling in `scripts/demo/sampling.py`.
117
+
We provide a [streamlit](https://streamlit.io/) demo for text-to-image and image-to-image sampling in `scripts/demo/sampling.py`.
118
+
We provide file hashes for the complete file as well as for only the saved tensors in the file (see [Model Spec](https://github.com/Stability-AI/ModelSpec) for a script to evaluate that).
@@ -134,13 +143,13 @@ The weights of SDXL-1.0 are available (subject to a [`CreativeML Open RAIL++-M`
134
143
135
144
**SDXL-0.9:**
136
145
The weights of SDXL-0.9 are available and subject to a [research license](model_licenses/LICENSE-SDXL0.9).
137
-
If you would like to access these models for your research, please apply using one of the following links:
138
-
[SDXL-base-0.9 model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-refiner-0.9](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
139
-
This means that you can apply for any of the two links - and if you are granted - you can access both.
146
+
If you would like to access these models for your research, please apply using one of the following links:
147
+
[SDXL-base-0.9 model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-refiner-0.9](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
148
+
This means that you can apply for any of the two links - and if you are granted - you can access both.
140
149
Please log in to your Hugging Face Account with your organization email to request access.
141
150
142
151
143
-
After obtaining the weights, place them into `checkpoints/`.
152
+
After obtaining the weights, place them into `checkpoints/`.
**NOTE 1:** Using the non-toy-dataset configs `configs/example_training/imagenet-f8_cond.yaml`, `configs/example_training/txt2img-clipl.yaml` and `configs/example_training/txt2img-clipl-legacy-ucg-training.yaml` for training will require edits depending on the used dataset (which is expected to stored in tar-file in the [webdataset-format](https://github.com/webdataset/webdataset)). To find the parts which have to be adapted, search for comments containing `USER:` in the respective config.
206
+
**NOTE 1:** Using the non-toy-dataset configs `configs/example_training/imagenet-f8_cond.yaml`, `configs/example_training/txt2img-clipl.yaml` and `configs/example_training/txt2img-clipl-legacy-ucg-training.yaml` for training will require edits depending on the used dataset (which is expected to stored in tar-file in the [webdataset-format](https://github.com/webdataset/webdataset)). To find the parts which have to be adapted, search for comments containing `USER:` in the respective config.
198
207
199
208
**NOTE 2:** This repository supports both `pytorch1.13` and `pytorch2`for training generative models. However for autoencoder training as e.g. in `configs/example_training/autoencoder/kl-f4/imagenet-attnfree-logvar.yaml`, only `pytorch1.13` is supported.
0 commit comments