SPADE: Spatial-Aware Denoising Network for Open-vocabulary Panoptic Scene Graph Generation with Long- and Local-range Context Reasoning

The IEEE/CVF International Conference on Computer Vision (ICCV) 2025
Xin Hu 1 , Ke Qin 12 , Guiduo Duan 12 , Ming Li 4 , Yuan-Fang Li 3 , Tao He 1,2,*
1 The Laboratory of Intelligent Collaborative Computing of UESTC
2 Ubiquitous Intelligence and Trusted Services Key Laboratory of Sichuan Province
3 Faculty of Information Technology, Monash University
4 Guangdong Laboratory of Artificial Intelligence and Digital Economy (SZ)
* Corresponding author

Abstract

Panoptic Scene Graph Generation (PSG) integrates instance segmentation with relation understanding to capture pixel-level structural relationships in complex scenes. Although recent approaches leveraging pre-trained vision-language models (VLMs) have significantly improved performance in the open-vocabulary setting, they commonly ignore the inherent limitations of VLMs in spatial relation reasoning, such as difficulty in distinguishing object relative positions, which results in suboptimal relation prediction. Motivated by the denoising diffusion model's inversion process in preserving the spatial structure of input images, we propose SPADE (SPatial-Aware Denoising-nEtwork) framework---a novel approach for open-vocabulary PSG. SPADE consists of two key steps: (1) inversion-guided calibration for the UNet adaptation, and (2) spatial-aware context reasoning. In the first step, we calibrate a general pre-trained teacher diffusion model into a PSG-specific denoising network with cross-attention maps derived during inversion through a lightweight LoRA-based fine-tuning strategy. In the second step, we develop a spatial-aware relation graph transformer that captures both local and long-range contextual information, facilitating the generation of high-quality relation queries. Extensive experiments on benchmark PSG and Visual Genome datasets demonstrate that SPADE outperforms state-of-the-art methods in both closed- and open-set scenarios, particularly for spatial relationship prediction.

Background

Panoptic Scene Graph Generation (PSG) is a powerful paradigm that unifies instance segmentation and relationship prediction to enable pixel-level, structured scene understanding. Unlike traditional Scene Graph Generation (SGG), which typically operates on detected bounding boxes and a closed vocabulary, PSG requires simultaneously segmenting all entities and modeling their interactions as subject-predicate-object triples.

In recent years, approaches based on Vision-Language Models (VLMs) have made significant progress in Open-vocabulary scenarios, and recent advances in Visual Language Models (VLMs) such as CLIP and BLIP have pushed the frontiers of Open-vocabulary comprehension, enabling models to recognize novel objects and relationships beyond fixed categories. . However, while these models perform well in semantic recognition, these approaches still have limitations in spatial relational reasoning, especially when it comes to accurately understanding relationships between distant objects pairs.

Motivation

Despite their impressive performance on open-world classification tasks, VLMs inherently lack spatial reasoning abilities because they are typically trained on image-caption datasets with limited spatial annotations. This limitation significantly affects their performance in predicting spatial predicates, especially when objects are far apart or exhibit subtle geometric relations.

Our investigation reveals a critical gap: ​​VLM-based PSG models underperform in spatial relation prediction​​, especially for distantly located objects (Fig. 1). Key observations include:

  • ​​​​Distance sensitivity​​: Recall@50 for spatial predicates drops by 6–10% when objects are >1/3 image width apart.

Inspired by the inversion process in denoising diffusion models — known for preserving fine-grained spatial structures in images — we asked: Can we integrate spatial knowledge into VLM-based frameworks without sacrificing their strong open-vocabulary capabilities?

Follow this insight, we adapt diffusion models to PSG via:

  • ​​Inversion-guided calibration​​ of cross-attention maps to inject spatial priors.
  • ​​Dual-context reasoning​​ combining long-range and local spatial cues.

    We hypothesize this synergy can overcome VLM limitations without sacrificing open-vocabulary generalization.

motivation

Methods

We propose SPADE (SPatial-Aware Denoising-nEtwork), a novel two-stage framework that explicitly enhances spatial reasoning while retaining open-vocabulary recognition capabilities.

framework

Inversion-Guided UNet Calibration In the first stage, we leverage a pre-trained diffusion model as a spatially-aware teacher. Using the inversion process, we extract cross-attention maps that serve as explicit spatial priors, guiding the adaptation of the UNet denoising backbone. To maintain the open-vocabulary recognition power, we adopt a lightweight fine-tuning strategy called LoRA (Low-Rank Adaptation), updating only a small set of parameters in cross-attention layers. This ensures that the pre-trained knowledge is preserved while injecting strong spatial cues.

gte

Spatial-Aware Context Reasoning In the second stage, we introduce a Spatial-Aware Relation Graph Transformer (RGT) to model both local and long-range context among segmented instances. By constructing a spatial-semantic graph where nodes represent instance masks and edges encode spatial and semantic affinities, the RGT iteratively refines object features through a combination of graph convolutions and self-attention mechanisms. This dual-context reasoning enables the model to capture subtle relationships that are crucial for accurate predicate prediction.

Citation

@article{SPADE,
  title={SPADE: Spatial-Aware Denoising Network for Open-vocabulary Panoptic Scene Graph Generation with Long- and Local-range Context Reasoning},
  author={Xin Hu, Ke Qin, Guiduo Duan, Ming Li, Yuanfang Li, Tao He},
  journal={iccv},
  year={2025}
}