Generative AI

GEBench: Benchmarking Image Generation Models as GUI Environments

HHaodong LiJJingwei WuQQuan SunGGuopeng LiJJuanxi TianHHuanyu ZhangYYanlin LaiRRuichuan AnHHongbo PengYYuhong DaiCChenxi LiCChunmei QingJJia WangZZiyang MengZZheng GeXXiangyu ZhangDDaxin Jiang
Published
February 9, 2026
Authors
17
Word Count
14,113

GEBench benchmarks image models as GUI simulators for autonomous agent training.

Abstract

Recent advancements in image generation models have enabled the prediction of future Graphical User Interface (GUI) states based on user instructions. However, existing benchmarks primarily focus on general domain visual fidelity, leaving the evaluation of state transitions and temporal coherence in GUI-specific contexts underexplored. To address this gap, we introduce GEBench, a comprehensive benchmark for evaluating dynamic interaction and temporal coherence in GUI generation. GEBench comprises 700 carefully curated samples spanning five task categories, covering both single-step interactions and multi-step trajectories across real-world and fictional scenarios, as well as grounding point localization. To support systematic evaluation, we propose GE-Score, a novel five-dimensional metric that assesses Goal Achievement, Interaction Logic, Content Consistency, UI Plausibility, and Visual Quality. Extensive evaluations on current models indicate that while they perform well on single-step transitions, they struggle significantly with maintaining temporal coherence and spatial grounding over longer interaction sequences. Our findings identify icon interpretation, text rendering, and localization precision as critical bottlenecks. This work provides a foundation for systematic assessment and suggests promising directions for future research toward building high-fidelity generative GUI environments. The code is available at: https://github.com/stepfun-ai/GEBench.

Key Takeaways

  • 1

    Image generation models excel at single steps but fail maintaining consistency across multiple GUI interactions.

  • 2

    Traditional benchmarks measure photorealism, not GUI simulation fidelity needed for autonomous agent training.

  • 3

    GEBench provides systematic evaluation across five task categories to test GUI generation reliability.

Limitations

  • Existing video generation benchmarks focus on continuous natural motion, not discrete GUI state changes.

  • Current image generation metrics like FID scores don't measure GUI consistency or text rendering accuracy.

Keywords

GUI generationtemporal coherencedynamic interactionvisual fidelityGUI-specific contextsGEBenchGE-Scoregoal achievementinteraction logiccontent consistencyUI plausibilityvisual quality

More in Generative AI

View all
GEBench: Benchmarking Image Generation Models as GUI Environments | Paperchime