Search

Intelligent Systems Lab @ PITT
Intelligent Systems Lab @ PITT
  • People
  • Publication
  • Dataset
  • Projects
    Inference-time Computation in Generative AI On-device AI Trustworthy AI Mobile and Connected Health Mobile and Edge Computing Systems Intelligent Wireless Systems
  • Video & Demos
  • Code
  • Contact

Trustworthy AI

The versatility of recent emerging AI techniques also brings challenges in ensuring the AI systems to be safe, fair, explainable, and to cause no harm. Our research aims at discovering potential malicious adaptations to AI models, and propose protections and mitigations against unwanted model usages.


Related Papers

FreezeAsGuard: Mitigating Illegal Adaptation of Diffusion Models via Selective Tensor Freezing
Illegally using fine-tuned diffusion models to forge human portraits has been a major threat to trustworthy AI. While most existing work focuses on detection of the AI-forged contents, our recent work instead aims to mitigate such illegal domain adaptation by applying safeguards on diffusion models. Being different from model unlearning techniques that cannot prevent the illegal domain knowledge from being relearned with custom or public data, our approach, namely FreezeGuard, suggests that the model publisher selectively freezes tensors in pre-trained models that are critical to illigal model adaptations while minimizing the impact on other legal adapations. Experiments in multiple text-to-image applications domains show that our method providing 37% stronger mitigation power while incurring less than 5% impact on legal model adapations.
Kai Huang, Haoming Wang, Wei Gao
December 2024 In arXiv
PDF Cite Code DOI
FreezeAsGuard: Mitigating Illegal Adaptation of Diffusion Models via Selective Tensor Freezing

© 2025 Pitt Intelligent Systems Lab.

Published with Wowchemy — the free, open source website builder that empowers creators.

Cite
Copy Download