| AnyEdit | AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea | arXiv | 2025 |
| ConsistI2V-Edit | Consistent Video Editing with Instruction-Tuned Diffusion Models | arXiv | 2025 |
| DiffusionPen | DiffusionPen: Towards Controllable Style-Specific Handwritten Text Generation | arXiv | 2025 |
| VACE | All-in-One Video Creation and Editing | Alibaba | 2025 |
| VideoPainter | Any-length Video Inpainting and Editing with Plug-and-Play Context Control | SIGGRAPH | 2025 |
| VideoGrain | Modulating Space-Time Attention for Multi-grained Video Editing | ICLR | 2025 |
| Señorita-2M | High-Quality Instruction-based Dataset for General Video Editing | arXiv | 2025 |
| MTV-Inpaint | Multi-Task Long Video Inpainting | arXiv | 2025 |
| MiniMax-Remover | Taming Bad Noise Helps Video Object Removal | arXiv | 2025 |
| LoRA-Edit | Controllable First-Frame-Guided Video Editing via Mask-Aware LoRA | arXiv | 2025 |
| VEGGIE | Instructional Editing and Reasoning of Video Concepts | arXiv | 2025 |
| StableV2V | Stablizing Shape Consistency in Video-to-Video Editing | arXiv | 2024 |
| AnyV2V | AnyV2V: A Tuning-Free Framework for Any Video-to-Video Editing Tasks | TMLR | 2024 |
| ReVideo | Remake a Video with Motion and Content Control | arXiv | 2024 |
| I2VEdit | First-Frame-Guided Video Editing via Image-to-Video Diffusion Models | arXiv | 2024 |
| FlowVid | Taming Imperfect Optical Flows for Consistent Video-to-Video Synthesis | arXiv | 2023 |
| TokenFlow | TokenFlow: Consistent Diffusion Features for Consistent Video Editing | ICLR | 2024 |
| Rerender A Video | Zero-Shot Text-Guided Video-to-Video Translation | SIGGRAPH Asia | 2023 |
| FateZero | Fusing Attentions for Zero-shot Text-based Video Editing | ICCV | 2023 |
| CoDeF | Content Deformation Fields for Temporally Consistent Video Processing | CVPR | 2024 |
| VideoSwap | Customized Video Subject Swapping with Interactive Semantic Point | CVPR | 2024 |
| FLATTEN | Optical Flow-guided Attention for Consistent T2V Editing | ICLR | 2024 |
| MotionEditor | Editing Video Motion via Content-Aware Diffusion | arXiv | 2023 |
| Ground-A-Video | Zero-shot Grounded Video Editing using T2I Diffusion Models | ICLR | 2024 |
| Tune-A-Video | One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation | ICCV | 2023 |
| Dreamix | Video Diffusion Models Are General Video Editors | Google | 2023 |
| Pix2video | Video Editing Using Image Diffusion | arXiv | 2023 |
| Video-P2P | Video Editing with Cross-attention Control | arXiv | 2023 |
| Edit-A-Video | Single Video Editing with Object-Aware Consistency | arXiv | 2023 |
| RAVE | RAVE: Randomized Noise Shuffling for Fast and Consistent Video Editing with Diffusion Models | CVPR | 2024 |
| MagicEdit | MagicEdit: High-Fidelity and Temporally Coherent Video Editing | arXiv | 2024 |