Trust Your Critic: Robust Reward Modeling and Reinforcement Learning for Faithful Image Editing and Generation

Xiangyu Zhao*, Peiyuan Zhang*, Junming Lin*, Tianhao Liang*, Yuchen Duan, Shengyuan Ding, Changyao Tian, Yuhang Zang, Junchi Yan, Xue Yang

Overview of the FIRM editing and generation data curation pipelines.

FIRM uses a difference-first editing pipeline and a plan-then-score generation pipeline to reduce critic hallucination and improve reward quality.

Abstract

Accurate critics are the bottleneck in RL for image editing and generation.

Reinforcement learning has become a promising tool for improving image editing and text-to-image generation, but current reward models often hallucinate, miss details, and assign noisy scores that misguide optimization. FIRM addresses this problem with tailored data curation pipelines, specialized reward models, a human-annotated evaluation benchmark, and reward formulations that better balance competing goals. The framework produces FIRM-Edit-8B and FIRM-Gen-8B, then uses them to guide RL for faithful editing and instruction-aligned generation.

Method

One framework, three connected pieces.

1. Better supervision for critics

FIRM-Edit scores editing with execution and consistency, while FIRM-Gen focuses on instruction following with explicit scoring plans. This yields cleaner reward data and more reliable critic behavior.

2. Dedicated evaluation benchmark

FIRM-Bench measures critic alignment with human judgment for both editing and generation, using balanced score distributions and controlled prompt difficulty.

3. Reward shaping for RL

CME for editing and QMA for generation use a Base-and-Bonus design, preventing easy reward hacking shortcuts and improving credit assignment during optimization.

Editing pipeline

Instead of asking a model to directly judge edited images end-to-end, FIRM first describes the visual differences between source and edited images, then scores execution and consistency from that structured evidence.

Generation pipeline

For text-to-image prompts, an LLM first expands the prompt into a checklist, and a multimodal evaluator scores the generated image against this plan dimension by dimension.

Reward design

CME makes execution a prerequisite for high editing reward, while QMA strengthens alignment with an explicit quality term to suppress low-quality but superficially compliant generations.

Reward formulas

CME

R = Execution * (0.6 + 0.4 * Consistency)

QMA

R = InsFollowing * (0.4 + 0.6 * Quality)

Both formulations are designed to stop the policy from maximizing the easiest sub-score while ignoring the real task objective.

Results

FIRM improves both critic quality and downstream RL performance.

Qualitative image editing results comparing reward models.

Editing guided by FIRM-Edit-8B

The paper highlights stronger edit execution while preserving task-irrelevant content, addressing the common failure mode of returning an almost unchanged image.

Qualitative text-to-image generation results comparing reward models.

Generation guided by FIRM-Gen-8B

Structured scoring plans improve instruction following, especially on prompts with multiple entities, styles, and spatial constraints.

BibTeX

@misc{zhao2026trustcriticrobustreward,
            title={Trust Your Critic: Robust Reward Modeling and Reinforcement Learning for Faithful Image Editing and Generation}, 
            author={Xiangyu Zhao and Peiyuan Zhang and Junming Lin and Tianhao Liang and Yuchen Duan and Shengyuan Ding and Changyao Tian and Yuhang Zang and Junchi Yan and Xue Yang},
            year={2026},
            eprint={2603.12247},
            archivePrefix={arXiv},
            primaryClass={cs.CV},
            url={https://arxiv.org/abs/2603.12247}, 
      }