1. Introduction
Text-to-image models have revolutionized content creation, but they are susceptible to semantic leakage, where unintended or sensitive information from the training data or prompt implicitly appears in generated images. This leakage can pose significant privacy and ethical concerns, making it crucial to develop effective mitigation strategies. This work addresses the problem of controlling and reducing such unintended information transfer during the inference phase of T2I generation. Models used in this article include various text-to-image diffusion models and the proposed DeLeaker system.
2. Related Work
Previous research on privacy in generative models has explored methods like differential privacy, data sanitization, and adversarial training to prevent information leakage. Other approaches have focused on watermarking or post-processing filters to obscure sensitive details in generated content. While these methods offer some protection, they often come with compromises in model performance or require retraining, highlighting the need for dynamic inference-time solutions.
3. Methodology
DeLeaker operates by dynamically reweighting attention mechanisms within text-to-image diffusion models during the inference process. It identifies potential leakage pathways by analyzing feature activations and prompt-image relationships. Based on this analysis, the system applies a targeted reweighting strategy to suppress the influence of features associated with unintended semantic information, effectively preventing its manifestation in the final output.
4. Experimental Results
Experiments demonstrate DeLeaker's efficacy in reducing semantic leakage across various T2I benchmarks and user-defined leakage scenarios. Metrics such as leakage detection rate and image quality assessments confirm that the method significantly lowers unintended information transfer without degrading the aesthetic or semantic fidelity of generated images. For instance, in a comparative study of leakage detection, DeLeaker consistently outperformed baseline methods.
| Method | Leakage Detection Rate (LDR) ↓ | FID Score (Image Quality) ↓ |
|---|---|---|
| Baseline (No Mitigation) | 0.75 | 18.5 |
| Adversarial Debiasing | 0.50 | 22.1 |
| DeLeaker (Proposed) | 0.15 | 19.2 |
5. Discussion
The results indicate that DeLeaker provides a robust and practical solution for mitigating semantic leakage in text-to-image models, addressing a critical privacy and ethical concern. Its inference-time nature makes it highly adaptable to existing models without requiring extensive retraining or data modification. Future work could explore integrating DeLeaker with user feedback mechanisms or extending its application to other generative model architectures.