Beyond Referring Expressions: Scenario Comprehension Visual Grounding

1Rice University  ·  2Johns Hopkins University  ·  3Northeastern University

Existing visual grounding benchmarks primarily evaluate alignment between image regions and literal referring expressions, where models can often succeed by matching a prominent named category. We explore a complementary and more challenging setting of scenario-based visual grounding, where the target must be inferred from roles, intentions, and relational context rather than explicit naming. We introduce Referring Scenario Comprehension (RSC), a benchmark designed for this setting. The queries in this benchmark are paragraph-length texts that describe object roles, user goals, and contextual cues that include deliberate references to distractor objects that often require deep understanding to resolve. Each instance is annotated with interpretable difficulty tags for uniqueness, clutter, size, overlap, and position which expose distinct failure modes and support fine-grained analysis. RSC contains approximately 31k training examples, 4k in-domain test examples, and a 3k out-of-distribution split with unseen object categories. We further propose ScenGround, a curriculum reasoning method serving as a reference point for this setting, combining supervised warm-starting with difficulty-aware reinforcement learning. Experiments show that scenario-based queries expose systematic failures in current models that standard benchmarks do not reveal, and that curriculum training improves performance on challenging slices and transfers to standard benchmarks.

arXiv Code Dataset Results

Overview

From Naming to Scenario Grounding

RSC Teaser figure
Figure 1: Referring Scenario Comprehension (RSC) vs. traditional referring expression comprehension (REC). Each row shows the same target object under both paradigms. Traditional REC queries often name the target category directly, allowing success via lexical matching. RSC instead pairs each image with a lengthy scenario-based query specifying a user role, goal, and multiple disambiguating cues — including explicit contrasts against competing objects. The RSC difficulty tags (U/C/S/O/P) characterize each instance, enabling fine-grained training and evaluation.

RSC Benchmark

Referring Scenario Comprehension Dataset

RSC replaces referring phrases with scenario-based queries that describe a user role, goal, and at least three disambiguating cues, and deliberately mentions competing objects to require deep understanding. Each instance is annotated with reasoning traces and five interpretable difficulty tags (Uniqueness, Clutter, Size, Overlap, and Position), which expose distinct failure modes and support fine-grained curriculum design and evaluation.

~31k
Training examples
4k
In-domain test
3k
OOD test (unseen)
52.7
Avg. query length (words)
9,086
Unique vocabulary tokens
RSC Pipeline
Figure 2: Phase 1 filters and balances source instances computing five interpretable difficulty tags to form a tag-balanced candidate pool. Phase 2 generates annotations via a two-stage process. Phase 3 applies automatic and human quality control. The final RSC dataset provides, per instance, a scenario query, reasoning traces, acceptable names, ground-truth box, and difficulty tags.
Query length distribution
Query length distribution. RSC queries peak around 50–60 words vs. under 10 for RefCOCO+/g.
Instance size distribution
Instance size distribution. RSC covers a broader range of target scales, with higher density at smaller instances.
Difficulty tag distributions
Difficulty Tag Distributions (RSC-ID vs. RSC-OOD). Per-tag marginals across all five axes: Uniqueness (U), Clutter (C), Size (S), Overlap (O), and Position (P). The ID split maintains near-balanced marginals by design; the OOD split skews toward non-unique (U2) and smaller instances, reflecting LVIS's fine-grained vocabulary.

ScenGround

Two-Stage Curriculum Reasoning

ScenGround is a two-stage curriculum reasoning method for scenario-based visual grounding. In Stage 1, Thought-Primed SFT (TP-SFT) aligns the model to the output schema and elicits faithful reasoning traces before a structured answer, using the easier RSC slices to stabilize interface learning. In Stage 2, Incentive-Curriculum GRPO (IC-GRPO) refines localization and disambiguation via shaped rewards coupling geometry (smooth IoU, center-consistency, out-of-bounds penalties) and alias-aware category rewards. The training follows a tag-aware curriculum, feeding more difficult non-unique, cluttered, overlapping, and off-center targets in the later stage. A prompt-template ensemble (PTE-8) further improves robustness across query surface forms.

ScenGround prompt schema
Figure 3: ScenGround prompt and output schema. The model reasons inside <think> and emits structured JSON inside <answer> with target_object and bbox [x,y,w,h]. Scenarios avoid category names and force disambiguation. IC-GRPO uses 8 prompt templates (PTE-8) for robustness.

Experiments

State-of-the-Art on RSC and Standard Benchmarks

Results reveal a consistent pattern: models with strong category accuracy tend to lag on localization, while strong detectors lack semantic reasoning. ScenGround substantially outperforms all baselines on ID mIoU and consistently reduces the localization–semantics trade-off across ID and OOD splits.

Model RSC In-Domain RSC Out-of-Domain
mIoU[email protected][email protected]Cat Acc mIoU[email protected][email protected]Cat Acc
Closed-source LLMs
GPT-4o 19.4113.235.3779.45 16.579.553.0862.00
Claude 3.7 16.648.323.7189.67 12.045.541.8758.98
Specialist Grounding Models (oracle settings ‡)
Grounding DINO (cat token) ‡ 44.6047.5542.03 32.1831.9927.89
Grounding DINO (ref. cue) ‡ 48.9951.8446.02 38.1238.2634.07
Open-source VLMs
InternVL2.5 8B 16.7611.886.7481.70 8.083.641.6136.50
Qwen3-VL 8B 15.4611.176.0575.04 7.383.701.4846.97
Qwen2.5-VL 7B 30.3127.4215.6630.86 21.5415.889.1920.82
ScenGround (Ours) 55.68 60.90 42.32 94.23 38.37 38.11 22.64 21.13

‡ Oracle settings: Grounding DINO receives privileged inputs (gold category name or short ref. cue) unavailable at inference.

Qualitative results
Figure 4: Qualitative results on RSC. ScenGround correctly localizes scenario-described targets — including distinguishing an illustrated animal on a book cover from a real dog in the foreground. Green = ground truth, blue/red = ScenGround prediction.
Top-20 categories
Top 20 most frequent categories in RSC-ID (79 COCO categories) and RSC-OOD (395 LVIS categories). OOD categories are disjoint at both string and synset level.

Citation

BibTeX

@article{he2026rsc,
  title = {Beyond Referring Expressions: Scenario Comprehension Visual Grounding},
  author = {He, Ruozhen and Shah, Nisarg A. and Dong, Qihua and Xiao, Zilin and Koo, Jaywon and Ordonez, Vicente},
  journal = {arXiv preprint},
  year = {2026},
  url = {https://arxiv.org/abs/XXXX.XXXXX}
}