CCEdit: Creative and Controllable Video Editing via Diffusion Models
Abstract
In this work, we present CCEdit, a versatile framework designed to address the challenges of creative and controllable video editing. CCEdit accommodates a wide spectrum of user editing requirements and enables enhanced creative control through an innovative approach that decouples video structure and appearance. We leverage the foundational ControlNet architecture to preserve structural integrity, while seamlessly integrating adaptable temporal modules compatible with state-of-the-art personalization techniques for text-to-image generation, such as DreamBooth and LoRA.Furthermore, we introduce reference-conditioned video editing, empowering users to exercise precise creative control over video editing through the more manageable process of editing key frames. Our extensive experimental evaluations confirm the exceptional functionality and editing capabilities of the proposed CCEdit framework. Demo video is available at https://www.youtube.com/watch?v=UQw4jq-igN4.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MagicEdit: High-Fidelity and Temporally Coherent Video Editing (2023)
- LAVIE: High-Quality Video Generation with Cascaded Latent Diffusion Models (2023)
- EVE: Efficient zero-shot text-based Video Editing with Depth Map Guidance and Temporal Consistency Constraints (2023)
- StableVideo: Text-driven Consistency-aware Diffusion Video Editing (2023)
- Edit Temporal-Consistent Videos with Image Diffusion Model (2023)
Please give a thumbs up to this comment if you found it helpful!
Models citing this paper 1
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper