--- tags: - text-to-image - layout-to-image - stable-diffusion - controlnet license: agpl-3.0 language: - en ---

Adversarial Supervision Makes Layout-to-Image Diffusion Models Thrive (ICLR 2024)

[**Project Page**](https://yumengli007.github.io/ALDM/) **|** [**ArXiv**](https://arxiv.org/abs/2401.08815) **|** [**Code**](https://github.com/boschresearch/ALDM)
This model repo contains checkpoints trained on Cityscapes and ADE20K datasets using methods proposed in ALDM. For usage instructions, please refer to our Github.
## Model information [ade20k_step9.ckpt](ade20k_step9.ckpt) and [cityscapes_step9.ckpt](cityscapes_step9.ckpt) are pretrained diffusion model weights for inference. [encoder_epoch_50.pth](encoder_epoch_50.pth), [decoder_epoch_50_20cls.pth](decoder_epoch_50_20cls.pth) and [decoder_epoch_50_151cls.pth](decoder_epoch_50_151cls.pth) are segmentation models used for discriminator intialization in training, which are adopted from pretrained UperNet101 [here](https://github.com/CSAILVision/semantic-segmentation-pytorch).