This repository contains the implementation of the following paper.
ProEdit: Inversion-based Editing From Prompts Done Right
Zhi Ouyang∗, Dian Zheng∗, Xiao-Ming Wu, Jian-Jian Jiang, Kun-Yu Lin, Jingke Meng+, Wei-Shi Zheng+
- 🔥 Updates
- 📣 Overview
- 📋 ToDo List
- 📖 Pipeline
- 🖼️ Code for Image Editing
- ✨ Text-driven Image / Video Editing
- 🎓 Editing by Instruction
- ✒️ Citation
♥️ Acknowledgement
- [2026.2.5] The code for Image Editing is released.
- [2025.12.28] The paper ProEdit is released on arXiv. 🚀
Overview of ProEdit. We propose a highly accurate, plug-and-play editing method for flow inversion that addresses the problem of excessive source image information injection, which prevents proper modification of attributes such as pose, number, and color. Our method has demonstrated impressive performance in both image editing and video editing tasks.
- Release the code for image editing
- Release the code for video editing
Pipeline of our ProEdit. The mask extraction module identifies the edited region based on source and target prompts during the first inversion step. After obtaining the inverted noise, we apply Latents-Shift to perturb the initial distribution in the edited region, reducing source image information. In selected sampling steps, we fuse source and target attention features in the edited region while directly injecting source features in non-edited regions to achieve accurate attribute editing and background preservation simultaneously.
For image editing, ProEdit employs FLUX as the backbone, and has been adapted to support four sampling solvers: Vanilla Flow, RF-Solver, Fireflow and UniEdit-Flow.
We have provided the code and demo for image editing using FLUX as the backbone, which can be found Here.
More results can be found in our project page.
With the assistance of a large language model, our method can directly perform edits guided by editing instructions.
If you find our repo useful for your research, please consider citing our paper:
@article{ouyang2025proedit,
title={ProEdit: Inversion-based Editing From Prompts Done Right},
author={Ouyang, Zhi and Zheng, Dian and Wu, Xiao-Ming and Jiang, Jian-Jian and Lin, Kun-Yu and Meng, Jingke and Zheng, Wei-Shi},
journal={arXiv preprint arXiv:2512.22118},
year={2025}
}We sincerely thank FireFlow, RF-Solver, UniEdit-Flow and FLUX for their awesome work! Additionally, we would also like to thank PnP-Inversion for providing comprehensive baseline survey and implementations, as well as their great benchmark.

