Recent advances in large-scale text-to-image diffusion models have heightened concerns about their
potential misuse,
especially in generating harmful or misleading content. This underscores the urgent need for effective
machine unlearning, i.e., removing specific knowledge or concepts from pretrained
models without compromising overall performance. One possible approach is Low-Rank Adaptation
(LoRA), which offers an efficient means to fine-tune models for targeted unlearning. However, LoRA
often inadvertently alters unrelated
content, leading to diminished image fidelity and realism. To address this limitation, we introduce UnGuide—a novel approach which incorporates UnGuidance, a dynamic inference
mechanism that leverages Classifier-Free Guidance (CFG) to exert precise control over the
unlearning process. UnGuide modulates the guidance scale based on the
stability of a few
first steps of denoising processes, enabling selective unlearning by LoRA adapter. For prompts containing
the erased concept, the LoRA module predominates and is counterbalanced
by the base model; for unrelated prompts, the base model governs generation, preserving content fidelity.
Empirical results demonstrate that UnGuide achieves controlled concept
removal and retains the expressive power of diffusion models, outperforming existing LoRA-based
methods in both object erasure and explicit content removal tasks.