Skip to content

Commit dea6059

Browse files
author
Jonas Müller
authored
Model hashes (#70)
* Added model hashes * Fix link
1 parent 299abbc commit dea6059

File tree

1 file changed

+24
-15
lines changed

1 file changed

+24
-15
lines changed

README.md

+24-15
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
## News
66

77
**July 26, 2023**
8-
- We are releasing two new open models with a permissive [`CreativeML Open RAIL++-M` license](model_licenses/LICENSE-SDXL1.0):
8+
- We are releasing two new open models with a permissive [`CreativeML Open RAIL++-M` license](model_licenses/LICENSE-SDXL1.0) (see [Inference](#inference) for file hashes):
99
- [SDXL-base-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0): An improved version over `SDXL-base-0.9`.
1010
- [SDXL-refiner-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0): An improved version over `SDXL-refiner-0.9`.
1111

@@ -22,11 +22,11 @@
2222
- `SDXL-base-0.9`: The base model was trained on a variety of aspect ratios on images with resolution 1024^2. The base model uses [OpenCLIP-ViT/G](https://github.com/mlfoundations/open_clip) and [CLIP-ViT/L](https://github.com/openai/CLIP/tree/main) for text encoding whereas the refiner model only uses the OpenCLIP model.
2323
- `SDXL-refiner-0.9`: The refiner has been trained to denoise small noise levels of high quality data and as such is not expected to work as a text-to-image model; instead, it should only be used as an image-to-image model.
2424

25-
If you would like to access these models for your research, please apply using one of the following links:
26-
[SDXL-0.9-Base model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-0.9-Refiner](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
27-
This means that you can apply for any of the two links - and if you are granted - you can access both.
25+
If you would like to access these models for your research, please apply using one of the following links:
26+
[SDXL-0.9-Base model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-0.9-Refiner](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
27+
This means that you can apply for any of the two links - and if you are granted - you can access both.
2828
Please log in to your Hugging Face Account with your organization email to request access.
29-
**We plan to do a full release soon (July).**
29+
**We plan to do a full release soon (July).**
3030

3131
## The codebase
3232

@@ -58,12 +58,12 @@ cd generative-models
5858

5959
#### 2. Setting up the virtualenv
6060

61-
This is assuming you have navigated to the `generative-models` root after cloning it.
61+
This is assuming you have navigated to the `generative-models` root after cloning it.
6262

6363
**NOTE:** This is tested under `python3.8` and `python3.10`. For other python versions, you might encounter version conflicts.
6464

6565

66-
**PyTorch 1.13**
66+
**PyTorch 1.13**
6767

6868
```shell
6969
# install required packages from pypi
@@ -72,7 +72,7 @@ source .pt13/bin/activate
7272
pip3 install -r requirements/pt13.txt
7373
```
7474

75-
**PyTorch 2.0**
75+
**PyTorch 2.0**
7676

7777

7878
```shell
@@ -112,13 +112,22 @@ You will find the built package in `dist/`. You can install the wheel with `pip
112112
Note that the package does **not** currently specify dependencies; you will need to install the required packages,
113113
depending on your use case and PyTorch version, manually.
114114

115-
## Inference:
115+
## Inference
116116

117-
We provide a [streamlit](https://streamlit.io/) demo for text-to-image and image-to-image sampling in `scripts/demo/sampling.py`.
117+
We provide a [streamlit](https://streamlit.io/) demo for text-to-image and image-to-image sampling in `scripts/demo/sampling.py`.
118+
We provide file hashes for the complete file as well as for only the saved tensors in the file (see [Model Spec](https://github.com/Stability-AI/ModelSpec) for a script to evaluate that).
118119
The following models are currently supported:
119120

120121
- [SDXL-base-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0)
122+
```
123+
File Hash (sha256): 31e35c80fc4829d14f90153f4c74cd59c90b779f6afe05a74cd6120b893f7e5b
124+
Tensordata Hash (sha256): 0xd7a9105a900fd52748f20725fe52fe52b507fd36bee4fc107b1550a26e6ee1d7
125+
```
121126
- [SDXL-refiner-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0)
127+
```
128+
File Hash (sha256): 7440042bbdc8a24813002c09b6b69b64dc90fded4472613437b7f55f9b7d9c5f
129+
Tensordata Hash (sha256): 0x1a77d21bebc4b4de78c474a90cb74dc0d2217caf4061971dbfa75ad406b75d81
130+
```
122131
- [SDXL-base-0.9](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9)
123132
- [SDXL-refiner-0.9](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9)
124133
- [SD-2.1-512](https://huggingface.co/stabilityai/stable-diffusion-2-1-base/blob/main/v2-1_512-ema-pruned.safetensors)
@@ -134,13 +143,13 @@ The weights of SDXL-1.0 are available (subject to a [`CreativeML Open RAIL++-M`
134143

135144
**SDXL-0.9:**
136145
The weights of SDXL-0.9 are available and subject to a [research license](model_licenses/LICENSE-SDXL0.9).
137-
If you would like to access these models for your research, please apply using one of the following links:
138-
[SDXL-base-0.9 model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-refiner-0.9](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
139-
This means that you can apply for any of the two links - and if you are granted - you can access both.
146+
If you would like to access these models for your research, please apply using one of the following links:
147+
[SDXL-base-0.9 model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9), and [SDXL-refiner-0.9](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9).
148+
This means that you can apply for any of the two links - and if you are granted - you can access both.
140149
Please log in to your Hugging Face Account with your organization email to request access.
141150

142151

143-
After obtaining the weights, place them into `checkpoints/`.
152+
After obtaining the weights, place them into `checkpoints/`.
144153
Next, start the demo using
145154

146155
```
@@ -194,7 +203,7 @@ run
194203
python main.py --base configs/example_training/toy/mnist_cond.yaml
195204
```
196205

197-
**NOTE 1:** Using the non-toy-dataset configs `configs/example_training/imagenet-f8_cond.yaml`, `configs/example_training/txt2img-clipl.yaml` and `configs/example_training/txt2img-clipl-legacy-ucg-training.yaml` for training will require edits depending on the used dataset (which is expected to stored in tar-file in the [webdataset-format](https://github.com/webdataset/webdataset)). To find the parts which have to be adapted, search for comments containing `USER:` in the respective config.
206+
**NOTE 1:** Using the non-toy-dataset configs `configs/example_training/imagenet-f8_cond.yaml`, `configs/example_training/txt2img-clipl.yaml` and `configs/example_training/txt2img-clipl-legacy-ucg-training.yaml` for training will require edits depending on the used dataset (which is expected to stored in tar-file in the [webdataset-format](https://github.com/webdataset/webdataset)). To find the parts which have to be adapted, search for comments containing `USER:` in the respective config.
198207

199208
**NOTE 2:** This repository supports both `pytorch1.13` and `pytorch2`for training generative models. However for autoencoder training as e.g. in `configs/example_training/autoencoder/kl-f4/imagenet-attnfree-logvar.yaml`, only `pytorch1.13` is supported.
200209

0 commit comments

Comments
 (0)