Consistent Text-to-Image Generation via Scene De-contextualization

Author A1,Author B2,Author C1

1affiliation A 2affiliation B

主图 11
主图 22
主图 33
主图 44

This is a sentence that introduces the research direction and significance of the entire paper.

Abstract

Consistent text-to-image (T2I) generation seeks to produce identity-preserving images of the same subject across diverse scenes, yet it often fails due to a phenomenon called identity (ID) shift. Previous methods have tackled this issue, but typically rely on the unrealistic assumption of knowing all target scenes in advance. This paper reveals that a key source of ID shift is the native correlation between subject and scene context, called scene contextualization, which arises naturally as T2I models fit the training distribution of vast natural images. We formally prove the near-universality of this scene-ID correlation and derive theoretical bounds on its strength. On this basis, we propose a novel, efficient, training-free prompt embedding editing approach, called Scene De-Contextualization (SDeC), that imposes an inversion process of T2I’s built-in scene contextualization. Specifically, it identifies and suppresses the latent scene-ID correlation within the ID prompt’s embedding by quantifying the SVD directional stability to adaptively re-weight the corresponding eigenvalues. Critically, SDeC allows for per-scene use (one scene per prompt) without requiring prior access to all target scenes. This makes it a highly flexible and general solution well-suited to real-world applications where such prior knowledge is often unavailable or varies over time. Experiments demonstrate that SDeC significantly enhances identity preservation while maintaining scene diversity.

Introduction

Briefly describe the research background, related work or challenges.

This paper investigates the problem of ID shift in text-to-image (T2I) generation—where the same subject’s appearance changes across different generated scenes. While T2I models like Stable Diffusion excel at visual fidelity, they struggle to maintain consistent identity in narrative or sequential tasks. The authors attribute this to scene contextualization, a bias caused by the model’s learned association between subjects and typical scene contexts. To address this, they introduce Scene De-Contextualization (SDeC), which mathematically reverses contextual bias through eigenvalue optimization using SVD. SDeC decouples identity embeddings from scene correlations, allowing consistent generation without needing full target-scene supervision. Experiments show that SDeC improves identity preservation and scene diversity across various models and tasks.

Visualization of how scene contextualization affects subject identity consistency in text-to-image generation.

背景图
Figure 1: Illustration of scene contextualization with SDXL.
Left: The attire of the subject varies with the site. Right: The subject’s clothing changes with the season.

Scene Contextualizaztion

Overview of the method framework, key modules and innovations.

We propose SDeC (Scene De-Contextualization), which mitigates scene influence by identifying and suppressing the latent scene–ID correlation subspace in ID embeddings. Using SVD-based optimization and eigenvalue modulation, SDeC enhances stable directions and reconstructs a refined identity representation, improving identity consistency in generated images.

示例图 1
Description

Further Analysis

Both the soft–estimation and absolute eigenvalue excursion designs in SDeC notably improve performance by enhancing representation accuracy and numerical stability.

示例图 1
Qualitative results. CS: Consistency

Experimental Details

Impact 01
Supplementary qualitative comparison results.
Impact 02
Results of baseline experiment where the images are arranged in a U-shape. All scenarios share the same ID prompt, and the scene prompts are different, as in the setting of this paper.
Impact 03
Results of experiment with major perturbation where the images are arranged in a U-shape.
The intermediate scenarios (#2∼#11) significantly differ from the first one.
Impact 04
Experiment with minor perturbation where the images are arranged in a U-shape. In the intermediate scenarios (#2∼#11), the scene prompts generated by ChatGPT-5.0 follow a literary style similar to that of the first, while the ID prompts remain consistent with the first, differing only in the substitution of the subject with a related concept.
Impact 05
Extended experiment with minor perturbations, with the number of images increased to the maximum continuous generation times of Nano Banana.

Poster

Display qualitative/quantitative results, supporting PDF preview and explanatory text.

Bibtex

@inproceedings{

}