Edit model card

SV3D-diffusers

This repo (https://github.com/chenguolin/sv3d-diffusers) provides scripts about:

  1. Spatio-temporal UNet (SV3DUNetSpatioTemporalConditionModel) and pipeline (StableVideo3DDiffusionPipeline) modified from SVD for SV3D in the diffusers convention.

  2. Converting the Stability-AI's SV3D-p UNet checkpoint to the diffusers convention.

  3. Infering the SV3D-p model with the diffusers library to synthesize a 21-frame orbital video around a 3D object from a single-view image (preprocessed by removing background and centering first).

Converted SV3D-p checkpoints have been uploaded to HuggingFaceπŸ€— chenguolin/sv3d-diffusers.

πŸš€ Usage

git clone https://github.com/chenguolin/sv3d-diffusers.git
# Please install PyTorch first according to your CUDA version
pip3 install -r requirements.txt
# If you can't access to HuggingFaceπŸ€—, try:
# export HF_ENDPOINT=https://hf-mirror.com
python3 infer.py --output_dir out/ --image_path assets/images/sculpture.png --elevation 10 --half_precision --seed -1

The synthesized video will save at out/ as a .gif file.

πŸ“Έ Results

Image preprocessing and random seed for different implementations are different, so the results are presented only for reference.

Implementation sculpture bag kunkun
SV3D-diffusers (Ours)
Official SV3D

πŸ“š Citation

If you find this repo helpful, please consider giving this repository a star 🌟 and citing the original SV3D paper.

@inproceedings{voleti2024sv3d,
   author={Voleti, Vikram and Yao, Chun-Han and Boss, Mark and Letts, Adam and Pankratz, David and Tochilkin,  Dmitrii and Laforte, Christian and Rombach, Robin and Jampani, Varun},
   title={{SV3D}: Novel Multi-view Synthesis and {3D} Generation from a Single Image using Latent Video Diffusion},
   booktitle={European Conference on Computer Vision (ECCV)},
   year={2024},
}
Downloads last month
22
Inference API
Inference API (serverless) does not yet support diffusers models for this pipeline type.

Dataset used to train chenguolin/sv3d-diffusers