pith. machine review for the scientific record. sign in

arxiv: 2604.09405 · v1 · submitted 2026-04-10 · 💻 cs.CV

Recognition: unknown

EGLOCE: Training-Free Energy-Guided Latent Optimization for Concept Erasure

Authors on Pith no claims yet

Pith reviewed 2026-05-10 16:46 UTC · model grok-4.3

classification 💻 cs.CV
keywords concept erasurediffusion modelstraining-freelatent optimizationenergy guidancetext-to-image generationinference-time methodssafe generation
0
0 comments X

The pith

EGLOCE removes specific concepts from diffusion model generations by optimizing latents at inference time using dual energy guidance.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

This paper presents a training-free method to erase unwanted concepts such as explicit content or copyrighted styles from images made by text-to-image diffusion models. It redirects the noisy latent during the sampling process by applying a repulsion energy that pushes away from the target concept and a retention energy that keeps the output faithful to the prompt. A sympathetic reader would care because existing approaches either demand expensive retraining or weaken the model in other ways, while this one promises safer generation that works on the fly with current systems. If the method holds, it could make concept control a simple addition rather than a full model overhaul.

Core claim

The central claim is that concept erasure succeeds through dual-objective energy-guided latent optimization performed entirely at inference. A repulsion energy term steers the latent away from the target concept via gradient descent, while a retention energy term maintains semantic alignment to the input prompt. This combination improves erasure over baselines that rely on altered weights or indirect guidance, preserves image quality and prompt fidelity, and holds up even against adversarial attacks.

What carries the argument

A dual-objective energy framework consisting of repulsion energy that steers the latent away from target concepts via gradient descent and retention energy that preserves prompt alignment, applied directly during inference sampling.

If this is right

  • Concept removal improves across existing baseline methods without retraining.
  • Image quality and prompt alignment stay intact after erasure.
  • Performance holds even when inputs include adversarial prompts.
  • The approach integrates directly with unmodified diffusion models.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same latent redirection could support on-the-fly user controls for other generation constraints beyond erasure.
  • Energy guidance might extend to selective feature addition or style transfer at inference.
  • Deployed systems could adopt per-prompt concept filters without maintaining multiple model versions.
  • This reduces dependence on upfront safety fine-tuning for new concepts.

Load-bearing premise

The repulsion and retention energies can be balanced during latent optimization without creating artifacts or prompt misalignment.

What would settle it

A set of generated images where the target concept still appears or where image quality and prompt match degrade noticeably compared to the baseline method.

Figures

Figures reproduced from arXiv: 2604.09405 by Junyeong Ahn, Seojin Yoon, Sungyong Baik.

Figure 1
Figure 1. Figure 1: Our proposed framework EGLOCE ensures that noise latents are both semantically aligned with the input prompt and remain [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of the EGLOCE framework. The flowchart illustrates how repulsion and retention energies are applied during diffusion [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Qualitative results of the proposed method. [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Qualitative results on nudity erasure across four baselines. [PITH_FULL_IMAGE:figures/full_fig_p008_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Failure cases show that, in some cases, integrating our method to the baseline brings little visible change over the original results, [PITH_FULL_IMAGE:figures/full_fig_p008_5.png] view at source ↗
read the original abstract

As text-to-image diffusion models grow increasingly prevalent, the ability to remove specific concepts-mostly explicit content and many copyrighted characters or styles-has become essential for safety and compliance. Existing unlearning approaches often require costly re-training, modify parameters at the cost of degradation of unrelated concept fidelity, or depend on indirect inference-time adjustment that compromise the effectiveness of concept erasure. Inspired by the success of energy-guided sampling for preservation of the condition of diffusion models, we introduce Energy-Guided Latent Optimization for Concept Erasure (EGLOCE), a training-free approach that removes unwanted concepts by re-directing noisy latent during inference. Our method employs a dual-objective framework: a repulsion energy that steers generation away from target concepts via gradient descent in latent space, and a retention energy that preserves semantic alignment to the original prompt. Combined with previous approaches that either require erroneous modified model weights or provide weak inference-time guidance, EGLOCE operates entirely at inference and enhances erasure performance, enabling plug-and-play integration. Extensive experiments demonstrate that EGLOCE improves concept removal while maintaining image quality and prompt alignment across baselines, even with adversarial attacks. To the best of our knowledge, our work is the first to establish a new paradigm for safe and controllable image generation through dual energy-based guidance during sampling.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper proposes EGLOCE, a training-free inference-time method for concept erasure in text-to-image diffusion models. It redirects noisy latents via gradient descent on a dual-objective energy: a repulsion term that steers away from a target concept and a retention term that preserves semantic alignment with the original prompt. The approach is presented as plug-and-play, compatible with existing baselines, and robust to adversarial attacks while maintaining image quality and prompt fidelity. Extensive experiments are claimed to support improved erasure over prior retraining-based and weak guidance methods.

Significance. If the dual-energy balancing proves robust with fixed hyperparameters and the claimed gains hold under quantitative evaluation, the work would provide a practical, zero-training alternative for controllable generation that avoids parameter modification and its side effects on unrelated concepts. The inference-only nature and explicit dual-objective formulation distinguish it from prior art and could enable safer deployment of diffusion models.

major comments (3)
  1. [§3] §3 (Method) and Algorithm 1: The repulsion and retention energies are described at a high level but lack explicit functional forms or the auxiliary model used to compute concept presence. Without these, it is impossible to assess whether the joint optimization is stable or whether the balancing weights are truly fixed across concepts rather than tuned per target.
  2. [§4] §4 (Experiments): The abstract asserts extensive experiments, robustness to adversarial attacks, and maintained prompt alignment, yet no quantitative metrics, baselines, or ablation on hyperparameter sensitivity are referenced. This leaves the central claim that the dual-objective framework avoids artifacts or per-concept tuning unsupported in the provided text.
  3. [§3.2] §3.2 (Energy balancing): The claim that repulsion and retention can be jointly optimized via gradient descent at inference without degrading quality hinges on the relative weighting and step count being generally robust. No evidence is given that these choices generalize without introducing prompt misalignment on unrelated concepts.
minor comments (2)
  1. [Abstract] The abstract states 'to the best of our knowledge' this is the first dual energy-based guidance paradigm; a brief related-work comparison table would clarify novelty relative to prior energy-guided sampling papers.
  2. [§3] Notation for the latent optimization update (e.g., the gradient step size and number of steps) should be introduced with a clear equation rather than prose description.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive feedback on our manuscript. We address each major comment point by point below and have revised the paper to provide the requested clarifications, explicit formulations, and additional experimental evidence.

read point-by-point responses
  1. Referee: [§3] §3 (Method) and Algorithm 1: The repulsion and retention energies are described at a high level but lack explicit functional forms or the auxiliary model used to compute concept presence. Without these, it is impossible to assess whether the joint optimization is stable or whether the balancing weights are truly fixed across concepts rather than tuned per target.

    Authors: We agree that the initial presentation was insufficiently detailed. In the revised manuscript, Section 3 now explicitly defines the repulsion energy as E_rep(z_t) = -log(1 - sim(CLIP(z_t), concept_embedding)) and the retention energy as E_ret(z_t) = -sim(CLIP(z_t), prompt_embedding), where the auxiliary model is a frozen CLIP ViT-L/14 encoder used for zero-shot concept detection via cosine similarity. The balancing weights are fixed at λ_rep = 1.0 and λ_ret = 0.5 for all concepts (no per-target tuning), as stated in the updated text and verified through cross-concept experiments. Algorithm 1 has been expanded with the exact gradient update equations and these constants. revision: yes

  2. Referee: [§4] §4 (Experiments): The abstract asserts extensive experiments, robustness to adversarial attacks, and maintained prompt alignment, yet no quantitative metrics, baselines, or ablation on hyperparameter sensitivity are referenced. This leaves the central claim that the dual-objective framework avoids artifacts or per-concept tuning unsupported in the provided text.

    Authors: We acknowledge the need for clearer referencing of results. The revised Section 4 now includes quantitative tables reporting CLIP-score for prompt alignment, concept classification accuracy for erasure success, and FID for image quality, with comparisons to baselines including ESD, UCE, and prior inference-time methods. We have added an ablation study on hyperparameter sensitivity (varying step count and weights) and results demonstrating robustness under adversarial prompts, confirming no per-concept tuning is required and that artifacts are avoided. revision: yes

  3. Referee: [§3.2] §3.2 (Energy balancing): The claim that repulsion and retention can be jointly optimized via gradient descent at inference without degrading quality hinges on the relative weighting and step count being generally robust. No evidence is given that these choices generalize without introducing prompt misalignment on unrelated concepts.

    Authors: We accept that additional evidence is warranted. The revised Section 3.2 now details the joint gradient descent procedure (10-20 steps with fixed weights) and includes new ablation results across 15 unrelated concepts and diverse prompts. These show that prompt alignment (CLIP similarity) remains stable with no measurable degradation on non-target concepts, supported by quantitative plots of the energy trade-off and qualitative examples. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation applies energy guidance directly without reduction to inputs or self-citations.

full rationale

The paper introduces EGLOCE as a training-free inference-time method that redirects noisy latents via a dual-objective energy framework (repulsion from target concepts plus retention of prompt semantics). No equations, fitted parameters, or self-citations appear in the provided text that would make any claimed result equivalent to its inputs by construction. The approach is presented as an application of existing energy-guided sampling ideas to concept erasure, with experiments claimed to validate performance; this does not trigger any of the enumerated circularity patterns. The derivation chain remains self-contained and does not rely on renaming, smuggling ansatzes, or load-bearing self-references.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The approach rests on the domain assumption that energy functions derived from the diffusion model can steer latents effectively, plus likely free parameters for balancing the two energies.

free parameters (1)
  • energy balancing weights
    Hyperparameters needed to trade off repulsion from target concept against prompt retention; not specified in abstract but required for the dual-objective framework.
axioms (1)
  • domain assumption Energy-guided sampling preserves conditioning in diffusion models
    Explicitly invoked as inspiration for the method.

pith-pipeline@v0.9.0 · 5531 in / 1178 out tokens · 54692 ms · 2026-05-10T16:46:55.805001+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

54 extracted references · 13 canonical work pages · 4 internal anchors

  1. [1]

    co / CompVis/stable-diffusion-v1-4, 2022

    Stable diffusion.https : / / huggingface . co / CompVis/stable-diffusion-v1-4, 2022. 1, 2, 3

  2. [2]

    Nudenet: Neural nets for nudity classification, detection and selective censoring

    P Bedapudi. Nudenet: Neural nets for nudity classification, detection and selective censoring. 2019. 7, 1

  3. [3]

    Con- trastive cfg: Improving cfg in diffusion models by con- trasting positive and negative concepts.arXiv preprint arXiv:2411.17077, 2024

    Jinho Chang, Hyungjin Chung, and Jong Chul Ye. Con- trastive cfg: Improving cfg in diffusion models by con- trasting positive and negative concepts.arXiv preprint arXiv:2411.17077, 2024. 2

  4. [4]

    Prompting4debugging: Red-teaming text-to-image diffusion models by finding problematic prompts.arXiv preprint arXiv:2309.06135, 2023

    Zhi-Yi Chin, Chieh-Ming Jiang, Ching-Chun Huang, Pin- Yu Chen, and Wei-Chen Chiu. Prompting4debugging: Red- teaming text-to-image diffusion models by finding problem- atic prompts.arXiv preprint arXiv:2309.06135, 2023. 7, 1

  5. [5]

    Diffusion Posterior Sampling for General Noisy Inverse Problems

    Hyungjin Chung, Jeongsol Kim, Michael T Mccann, Marc L Klasky, and Jong Chul Ye. Diffusion posterior sam- pling for general noisy inverse problems.arXiv preprint arXiv:2209.14687, 2022. 3

  6. [6]

    Imagenet: A large-scale hierarchical image database

    Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009. 9

  7. [7]

    Diffusion models beat gans on image synthesis.Advances in neural informa- tion processing systems, 34:8780–8794, 2021

    Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis.Advances in neural informa- tion processing systems, 34:8780–8794, 2021. 1

  8. [8]

    Implicit generation and mod- eling with energy based models.Advances in neural infor- mation processing systems, 32, 2019

    Yilun Du and Igor Mordatch. Implicit generation and mod- eling with energy based models.Advances in neural infor- mation processing systems, 32, 2019. 2

  9. [9]

    Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc

    Yilun Du, Conor Durkan, Robin Strudel, Joshua B Tenen- baum, Sander Dieleman, Rob Fergus, Jascha Sohl-Dickstein, Arnaud Doucet, and Will Sussman Grathwohl. Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc. InInternational conference on machine learning, pages 8489–8510. PMLR, 2023. 2

  10. [10]

    Tweedie’s formula and selection bias.Jour- nal of the American Statistical Association, 106(496):1602– 1614, 2011

    Bradley Efron. Tweedie’s formula and selection bias.Jour- nal of the American Statistical Association, 106(496):1602– 1614, 2011. 3

  11. [11]

    Zero-residual concept erasure via progres- sive alignment in text-to-image models

    Instance Erasure. Zero-residual concept erasure via progres- sive alignment in text-to-image models. 2

  12. [12]

    arXiv preprint arXiv:2310.12508 (2023)

    Chongyu Fan, Jiancheng Liu, Yihua Zhang, Eric Wong, Den- nis Wei, and Sijia Liu. Salun: Empowering machine unlearn- ing via gradient-based weight saliency in both image classi- fication and generation.arXiv preprint arXiv:2310.12508,

  13. [13]

    Erasing concepts from diffusion models

    Rohit Gandikota, Joanna Materzynska, Jaden Fiotto- Kaufman, and David Bau. Erasing concepts from diffusion models. InProceedings of the IEEE/CVF international con- ference on computer vision, pages 2426–2436, 2023. 2, 4, 5, 6, 1

  14. [14]

    Unified concept editing in diffusion models

    Rohit Gandikota, Hadas Orgad, Yonatan Belinkov, Joanna Materzy´nska, and David Bau. Unified concept editing in diffusion models. InProceedings of the IEEE/CVF Win- ter Conference on Applications of Computer Vision, pages 5111–5120, 2024. 4

  15. [15]

    Reliable and efficient concept erasure of text-to- image diffusion models

    Chao Gong, Kai Chen, Zhipeng Wei, Jingjing Chen, and Yu- Gang Jiang. Reliable and efficient concept erasure of text-to- image diffusion models. InEuropean Conference on Com- puter Vision, pages 73–88. Springer, 2024. 2, 5, 6, 1

  16. [16]

    Deep residual learning for image recognition

    Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. InProceed- ings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016. 9

  17. [17]

    Selective amnesia: A contin- ual learning approach to forgetting in deep generative mod- els.Advances in Neural Information Processing Systems, 36,

    Alvin Heng and Harold Soh. Selective amnesia: A contin- ual learning approach to forgetting in deep generative mod- els.Advances in Neural Information Processing Systems, 36,

  18. [18]

    Gans trained by a two time-scale update rule converge to a local nash equilib- rium.Advances in neural information processing systems, 30, 2017

    Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilib- rium.Advances in neural information processing systems, 30, 2017. 7, 1

  19. [19]

    Training products of experts by mini- mizing contrastive divergence.Neural computation, 14(8): 1771–1800, 2002

    Geoffrey E Hinton. Training products of experts by mini- mizing contrastive divergence.Neural computation, 14(8): 1771–1800, 2002. 2

  20. [20]

    Classifier-Free Diffusion Guidance

    Jonathan Ho and Tim Salimans. Classifier-free diffusion guidance.arXiv preprint arXiv:2207.12598, 2022. 1, 3

  21. [21]

    Denoising dif- fusion probabilistic models.Advances in neural information processing systems, 33:6840–6851, 2020

    Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising dif- fusion probabilistic models.Advances in neural information processing systems, 33:6840–6851, 2020. 1

  22. [22]

    Re- celer: Reliable concept erasing of text-to-image diffusion models via lightweight erasers

    Chi-Pin Huang, Kai-Po Chang, Chung-Ting Tsai, Yung- Hsuan Lai, Fu-En Yang, and Yu-Chiang Frank Wang. Re- celer: Reliable concept erasing of text-to-image diffusion models via lightweight erasers. InEuropean Conference on Computer Vision, pages 360–376. Springer, 2024. 2

  23. [23]

    Training-free safe denoisers for safe use of diffusion models.arXiv preprint arXiv:2502.08011, 2025

    Mingyu Kim, Dongjun Kim, Amman Yusuf, Stefano Ermon, and Mijung Park. Training-free safe denoisers for safe use of diffusion models.arXiv preprint arXiv:2502.08011, 2025. 2

  24. [24]

    Identity-preserving distillation sampling by fixed-point iterator

    SeonHwa Kim, Jiwon Kim, Soobin Park, Donghoon Ahn, Jiwon Kang, Seungryong Kim, Kyong Hwan Jin, and Eunju Cha. Identity-preserving distillation sampling by fixed-point iterator. InProceedings of the Computer Vision and Pattern Recognition Conference, pages 11115–11124, 2025. 3, 5

  25. [25]

    Auto-Encoding Variational Bayes

    Diederik P Kingma and Max Welling. Auto-encoding varia- tional bayes.arXiv preprint arXiv:1312.6114, 2013. 3

  26. [26]

    Koulischer, J

    Felix Koulischer, Johannes Deleu, Gabriel Raya, Thomas Demeester, and Luca Ambrogioni. Dynamic negative guid- ance of diffusion models.arXiv preprint arXiv:2410.14398,

  27. [27]

    Ablating con- cepts in text-to-image diffusion models

    Nupur Kumari, Bingliang Zhang, Sheng-Yu Wang, Eli Shechtman, Richard Zhang, and Jun-Yan Zhu. Ablating con- cepts in text-to-image diffusion models. InProceedings of the IEEE/CVF International Conference on Computer Vi- sion, pages 22691–22702, 2023. 2

  28. [28]

    A tutorial on energy-based learning.Predicting structured data, 1(0), 2006

    Yann LeCun, Sumit Chopra, Raia Hadsell, M Ranzato, Fujie Huang, et al. A tutorial on energy-based learning.Predicting structured data, 1(0), 2006. 3

  29. [29]

    Self-discovering interpretable diffusion latent di- rections for responsible text-to-image generation

    Hang Li, Chengzhi Shen, Philip Torr, V olker Tresp, and Jin- dong Gu. Self-discovering interpretable diffusion latent di- rections for responsible text-to-image generation. InPro- ceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12006–12016, 2024. 2

  30. [30]

    Speed: Scal- able, precise, and efficient concept erasure for diffusion mod- els.arXiv preprint arXiv:2503.07392, 2025

    Ouxiang Li, Yuan Wang, Xinting Hu, Houcheng Jiang, Tao Liang, Yanbin Hao, Guojun Ma, and Fuli Feng. Speed: Scal- able, precise, and efficient concept erasure for diffusion mod- els.arXiv preprint arXiv:2503.07392, 2025. 2, 4

  31. [31]

    Microsoft coco: Common objects in context

    Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll´ar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European conference on computer vision, pages 740–755. Springer, 2014. 7, 1

  32. [32]

    Dpm-solver: A fast ode solver for diffu- sion probabilistic model sampling in around 10 steps.Ad- vances in neural information processing systems, 35:5775– 5787, 2022

    Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver: A fast ode solver for diffu- sion probabilistic model sampling in around 10 steps.Ad- vances in neural information processing systems, 35:5775– 5787, 2022. 6

  33. [33]

    Mace: Mass concept erasure in diffu- sion models

    Shilin Lu, Zilan Wang, Leyang Li, Yanzhu Liu, and Adams Wai-Kin Kong. Mace: Mass concept erasure in diffu- sion models. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6430– 6440, 2024. 2

  34. [34]

    One-dimensional adapter to rule them all: Concepts diffusion models and erasing applications

    Mengyao Lyu, Yuhong Yang, Haiwen Hong, Hui Chen, Xuan Jin, Yuan He, Hui Xue, Jungong Han, and Guiguang Ding. One-dimensional adapter to rule them all: Concepts diffusion models and erasing applications. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7559–7568, 2024. 2

  35. [35]

    Improved denoising diffusion probabilistic models

    Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. InInternational conference on machine learning, pages 8162–8171. PMLR,

  36. [36]

    Gpt-5.1.https://openai.com/index/ gpt-5-1/, 2025

    OpenAI. Gpt-5.1.https://openai.com/index/ gpt-5-1/, 2025. Accessed: 2025-11-14. 9

  37. [37]

    Safecfg: Controlling harmful features with dynamic safe guidance for safe gen- eration.arXiv preprint arXiv:2412.16039, 2024

    Jiadong Pan, Liang Li, Hongcheng Gao, Zheng-Jun Zha, Qingming Huang, and Jiebo Luo. Safecfg: Controlling harmful features with dynamic safe guidance for safe gen- eration.arXiv preprint arXiv:2412.16039, 2024. 2

  38. [38]

    Proximal algorithms

    Neal Parikh, Stephen Boyd, et al. Proximal algorithms. Foundations and trends® in Optimization, 1(3):127–239,

  39. [39]

    Learning transferable visual models from natural language supervi- sion

    Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervi- sion. InInternational conference on machine learning, pages 8748–8763. PmLR, 2021. 4, 7, 1

  40. [40]

    High-resolution image synthesis with latent diffusion models

    Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Bj ¨orn Ommer. High-resolution image synthesis with latent diffusion models. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10684–10695, 2022. 3

  41. [41]

    Safe latent diffusion: Mitigating inappro- priate degeneration in diffusion models

    Patrick Schramowski, Manuel Brack, Bj ¨orn Deiseroth, and Kristian Kersting. Safe latent diffusion: Mitigating inappro- priate degeneration in diffusion models. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 22522–22531, 2023. 2, 5, 6, 7, 1

  42. [42]

    Denoising Diffusion Implicit Models

    Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models.arXiv preprint arXiv:2010.02502, 2020. 1

  43. [43]

    Ring-a-bell! how reliable are concept removal methods for diffusion models?arXiv preprint arXiv:2310.10012, 2023

    Yu-Lin Tsai, Chia-Yi Hsu, Chulin Xie, Chih-Hsun Lin, Jia- You Chen, Bo Li, Pin-Yu Chen, Chia-Mu Yu, and Chun-Ying Huang. Ring-a-bell! how reliable are concept removal meth- ods for diffusion models?arXiv preprint arXiv:2310.10012,

  44. [44]

    Mma-diffusion: Multimodal attack on diffusion models

    Yijun Yang, Ruiyuan Gao, Xiaosen Wang, Tsung-Yi Ho, Nan Xu, and Qiang Xu. Mma-diffusion: Multimodal attack on diffusion models. InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition, pages 7737–7746, 2024. 7, 1

  45. [45]

    Safree: Training-free and adaptive guard for safe text-to-image and video generation.arXiv preprint arXiv:2410.12761, 2024

    Jaehong Yoon, Shoubin Yu, Vaidehi Patil, Huaxiu Yao, and Mohit Bansal. Safree: Training-free and adaptive guard for safe text-to-image and video generation.arXiv preprint arXiv:2410.12761, 2024. 2, 5, 6, 1

  46. [46]

    Freedom: Training-free energy-guided condi- tional diffusion model

    Jiwen Yu, Yinhuai Wang, Chen Zhao, Bernard Ghanem, and Jian Zhang. Freedom: Training-free energy-guided condi- tional diffusion model. InProceedings of the IEEE/CVF In- ternational Conference on Computer Vision, pages 23174– 23184, 2023. 2, 3

  47. [47]

    Forget-me-not: Learning to for- get in text-to-image diffusion models

    Gong Zhang, Kai Wang, Xingqian Xu, Zhangyang Wang, and Humphrey Shi. Forget-me-not: Learning to for- get in text-to-image diffusion models. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 1755–1764, 2024. 2

  48. [48]

    Adding conditional control to text-to-image diffusion models

    Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF international conference on computer vision, pages 3836–3847, 2023. 1

  49. [49]

    The unreasonable effectiveness of deep features as a perceptual metric

    Richard Zhang, Phillip Isola, Alexei A Efros, Eli Shecht- man, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric. InProceedings of the IEEE conference on computer vision and pattern recogni- tion, pages 586–595, 2018. 8

  50. [50]

    Defensive unlearning with adversarial training for robust concept erasure in diffusion models.Advances in neu- ral information processing systems, 37:36748–36776, 2024

    Yimeng Zhang, Xin Chen, Jinghan Jia, Yihua Zhang, Chongyu Fan, Jiancheng Liu, Mingyi Hong, Ke Ding, and Sijia Liu. Defensive unlearning with adversarial training for robust concept erasure in diffusion models.Advances in neu- ral information processing systems, 37:36748–36776, 2024. 4

  51. [51]

    To gener- ate or not? safety-driven unlearned diffusion models are still easy to generate unsafe images

    Yimeng Zhang, Jinghan Jia, Xin Chen, Aochuan Chen, Yi- hua Zhang, Jiancheng Liu, Ke Ding, and Sijia Liu. To gener- ate or not? safety-driven unlearned diffusion models are still easy to generate unsafe images... for now. InEuropean Con- ference on Computer Vision, pages 385–403. Springer, 2024. 7, 1

  52. [52]

    Egsde: Unpaired image-to-image translation via energy-guided stochastic differential equations.Advances in Neural Infor- mation Processing Systems, 35:3609–3623, 2022

    Min Zhao, Fan Bao, Chongxuan Li, and Jun Zhu. Egsde: Unpaired image-to-image translation via energy-guided stochastic differential equations.Advances in Neural Infor- mation Processing Systems, 35:3609–3623, 2022. 2, 3 EGLOCE: Training-Free Energy-Guided Latent Optimization for Concept Erasure Supplementary Material In this supplementary material, we prov...

  53. [53]

    + ours setup which shows the best performance across baseline combinations, using nudity removal as the eval- uation task. We generate images under five adversar- ial prompt attacks: I2P [41], P4D [4], Ring-A-Bell [43], MMA-Diffusion [44], and UnlearnDiffAtk [51], and com- pute the attack success rate (ASR) using NudeNet [2]. To evaluate preservation of n...

  54. [54]

    and CLIP [39] on COCO-30k [31], where lower FID and higher CLIP indicate better visual quality and semantic alignment. B.1. Energy Guidance Application Timesteps In this section, we varyt start andt end, the time range over which our method is applied, to analyze how the behavior changes and to justify our current choice. In the default setting, we uset s...