You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
`indir` should contain images `*.png` and masks `<image_fname>_mask.png` like
117
107
the examples provided in `data/inpainting_examples`.
118
108
109
+
110
+
# Train your own LDMs
111
+
112
+
## Data preparation
113
+
114
+
### Faces
115
+
For downloading the CelebA-HQ and FFHQ datasets, proceed as described in the [taming-transformers](https://github.com/CompVis/taming-transformers#celeba-hq)
116
+
repository.
117
+
118
+
### LSUN
119
+
120
+
The LSUN datasets can be conveniently downloaded via the script available [here](https://github.com/fyu/lsun).
121
+
We performed a custom split into training and validation images, and provide the corresponding filenames
122
+
at [https://ommer-lab.com/files/lsun.zip](https://ommer-lab.com/files/lsun.zip).
123
+
After downloading, extract them to `./data/lsun`. The beds/cats/churches subsets should
124
+
also be placed/symlinked at `./data/lsun/bedrooms`/`./data/lsun/cats`/`./data/lsun/churches`, respectively.
125
+
126
+
### ImageNet
127
+
The code will try to download (through [Academic
128
+
Torrents](http://academictorrents.com/)) and prepare ImageNet the first time it
129
+
is used. However, since ImageNet is quite large, this requires a lot of disk
130
+
space and time. If you already have ImageNet on your disk, you can speed things
131
+
up by putting the data into
132
+
`${XDG_CACHE}/autoencoders/data/ILSVRC2012_{split}/data/` (which defaults to
133
+
`~/.cache/autoencoders/data/ILSVRC2012_{split}/data/`), where `{split}` is one
134
+
of `train`/`validation`. It should have the following structure:
where ``<config_spec>`` is one of {`celebahq-ldm-vq-4`(f=4, VQ-reg. autoencoder, spatial size 64x64x3),`ffhq-ldm-vq-4`(f=4, VQ-reg. autoencoder, spatial size 64x64x3),
0 commit comments