Generative AI

FireRed-Image-Edit-1.0 Techinical Report

SSuper Intelligence TeamCChanghao QiaoCChao HuiCChen LiCCunzheng WangDDejia SongJJiale ZhangJJing LiQQiang XiangRRunqi WangSShuang SunWWei ZhuXXu TangYYao HuYYibo ChenYYuhao HuangYYuxuan DuanZZhiyi ChenZZiyuan Guo
Published
February 12, 2026
Authors
19
Word Count
14,525

FireRed achieves production-ready image editing through data curation, not model scale.

Abstract

We present FireRed-Image-Edit, a diffusion transformer for instruction-based image editing that achieves state-of-the-art performance through systematic optimization of data curation, training methodology, and evaluation design. We construct a 1.6B-sample training corpus, comprising 900M text-to-image and 700M image editing pairs from diverse sources. After rigorous cleaning, stratification, auto-labeling, and two-stage filtering, we retain over 100M high-quality samples balanced between generation and editing, ensuring strong semantic coverage and instruction alignment. Our multi-stage training pipeline progressively builds editing capability via pre-training, supervised fine-tuning, and reinforcement learning. To improve data efficiency, we introduce a Multi-Condition Aware Bucket Sampler for variable-resolution batching and Stochastic Instruction Alignment with dynamic prompt re-indexing. To stabilize optimization and enhance controllability, we propose Asymmetric Gradient Optimization for DPO, DiffusionNFT with layout-aware OCR rewards for text editing, and a differentiable Consistency Loss for identity preservation. We further establish REDEdit-Bench, a comprehensive benchmark spanning 15 editing categories, including newly introduced beautification and low-level enhancement tasks. Extensive experiments on REDEdit-Bench and public benchmarks (ImgEdit and GEdit) demonstrate competitive or superior performance against both open-source and proprietary systems. We release code, models, and the benchmark suite to support future research.

Key Takeaways

  • 1

    State-of-the-art image editing comes from rigorous data curation and training optimization, not just larger models.

  • 2

    Hierarchical multi-stage filtering reduced 1.6 billion raw samples to 100 million high-quality training examples.

  • 3

    Multi-Condition Aware Bucket Sampler minimizes wasted computation by grouping images by aspect ratio and input count.

Limitations

  • Raw data at billion-scale is mostly garbage, requiring obsessive 94% filtering to achieve quality.

  • The script is incomplete, cutting off mid-sentence during the training methodology discussion.

Keywords

diffusion transformerdata curationtraining methodologyevaluation designtext-to-imageimage editingmulti-condition aware bucket samplerstochastic instruction alignmentasymmetric gradient optimizationDPOdiffusionNFTlayout-aware OCR rewardsdifferentiable consistency lossREDEdit-BenchImgEditGEdit

More in Generative AI

View all
FireRed-Image-Edit-1.0 Techinical Report | Paperchime