Yixin Liu (刘奕鑫)

4th-year CSE Ph.D. student at Lehigh University (Advisor: Prof. Lichao Sun)
Research: Generative AI Safety & Authenticity
Industry: Dolby Labs (10 months), Samsung Research America (11 months)
B.E. Software Engineering, South China University of Technology (2022)

News & Highlights
  • [Seeking Position] I am actively looking for full-time research scientist/engineer positions starting May 2026. Please reach out if you have opportunities!
  • [2025.12] Our DiffShortcut is accepted by KDD'26! An empirical framework rethinking shortcut learning in personalized diffusion model fine-tuning, and proposing a decoupled learning approach to defend against protective perturbations.
  • [2025.01] Our XAttnMark is accepted by ICML'25! State-of-the-art neural audio watermarking achieving joint detection and attribution. [Virtual Poster]
  • [2024.06] Presented MetaCloak as CVPR'24 Oral - watch the talk!
  • [2024.05] Our FViT is accepted by ICML'24 as Spotlight!
  • [2023.12] Our Stable Unlearnable Example (SEM) is accepted by AAAI'24! Achieving 3.91× speedup with improved protection efficacy.
Research Interest

My research focuses on Trustworthy AI – making foundation models safer and more reliable – with an emphasis on model vulnerabilities, data privacy, and content provenance. Grounded in adversarial learning and shortcut vulnerability, I think about the ML lifecycle in three stages: model internals, inputs, and outputs.

A Unified Framework for Fortifying the AI Lifecycle
[Details] A Lifecycle Approach to Trustworthy AI
  • Model Internals: Robust Explainability [SEAT (AAAI'23 Oral), FViTs (ICML'24 Spotlight)]: I study how attention mechanisms can be fragile and unfaithful when reused for explanations. Using adversarial latent-space training and diffusion-style denoising, I stabilize attention so that attention scores better reflect the model's true decision process and can be used more reliably as explanations for foundation models.
  • Input Side: Training Dynamics & Data Protection [MetaCloak (CVPR'24 Oral), DiffShortcut (KDD'26), SEM (AAAI'24), EditShield (ECCV'24), MUE (ICML'24 Workshop), GraphCloak, Linear Solver Analysis]: I study training dynamics and shortcut learning, showing that models often rely on brittle shortcuts. I use this understanding in two ways: (1) designing privacy-preserving perturbations that protect user data from unauthorized training, and (2) proposing decoupled training frameworks where diffusion models remain robust even when training data is intentionally corrupted.
  • Output Side: Content Provenance & Watermarking [XAttnMark (ICML'25), TextMarker]: I work on provenance for AI-generated content, addressing whether watermark signals can survive strong distortions, especially repeated generative editing. My watermarking architecture, combined with adversarial augmentation during training, can reliably detect AI-generated content after aggressive edits, helping maintain integrity of the overall ecosystem.
Professional Experience
  • Dolby Labs - Research Intern (Sep 2024 - Apr 2025, May 2025 - Aug 2025)
    Working on robust audio watermarking for content protection with Universal Music Group. Developed XAttnMark achieving state-of-the-art detection and attribution performance.
  • Samsung Research America - Research Intern (May 2024 - Aug 2024)
    Developed graph-based RAG system for log analysis, achieving +16 comprehensiveness score improvement. Also worked on DiffShortcut for defending protective perturbations in diffusion models.
  • Samsung Research America - Research Intern (May 2023 - Nov 2023)
    Proposed efficient defensive perturbation generation methods for data protection against diffusion models, resulting in MetaCloak (CVPR'24 Oral) and GraphCloak for graph data protection.
  • Lehigh University - Teaching Assistant
    CSE 017 Java Programming (Spring 2023), CSE 007 Python Programming (Spring 2024)
Invited Talks & Presentations
  • ICML 2025 Poster - "XAttnMark: Learning Robust Audio Watermarking with Cross-Attention" [Virtual Poster]
  • Dolby Lab Tech Summit - "Robust Audio Watermarking for the Music Industry" (June 2025) [Slides]
  • Microsoft ASG Research Day - "Adversarial Perturbation in Personalized Diffusion" (invited by Dr. Tianyi Chen, July 2024) [Slides]
  • CVPR 2024 Oral - "MetaCloak: Preventing Unauthorized T2I Diffusion Synthesis" (June 2024) [Video] [Slides]
Reviewer Service

NeurIPS'23'24, KDD'23'25, CVPR'24'25, ICML'24'25, ECCV'24 (Outstanding Reviewer), ICLR'25, ICASSP'25, IEEE TIP

Publications ( show selected / show all by topic / show all by date )

Topics: Unauthorized Exploitation / NLP Safety / Explainable AI / Model Compresssion / Applications (*/†: indicates equal contribution.)

XAttnMark: Learning Robust Audio Watermarking with Cross-Attention
Yixin Liu, Lie Lu, Jihui Jin, Lichao Sun, Andrea Fanelli

[Project Page] [Paper] [ICML Talk] ICML 2025

MetaCloak: Preventing Unauthorized Subject-driven Text-to-image Diffusion-based Synthesis via Meta-learning
Yixin Liu, Chenrui Fan, Yutong Dai, Xun Chen, Pan Zhou, Lichao Sun

[CVPR 2024 Oral]

Medical Unlearnable Examples: Securing Medical Data from Unauthorized Traning via Sparsity-Aware Local Masking
Weixiang Sun, Yixin Liu, Zhiling Yan, Kaidi Xu, Lichao Sun

[ICML'24 Next Gen AI Safety 2024 Workshop]

Stable Unlearnable Example: Enhancing the Robustness of Unlearnable Examples via Stable Error-Minimizing Noise
Yixin Liu, Kaidi Xu, Xun Chen, Lichao Sun

[AAAI 2024]

Improving Faithfulness for Vision Transformers
Lijie Hu*, Yixin Liu*, Ninghao Liu, Mengdi Huai, Lichao Sun and Di Wang

[ICML 2024 Spotlight]

GraphCloak: Safeguarding Graph-structured Data from Unauthorized Exploitation
Yixin Liu, Chenrui Fan, Xun Chen, Pan Zhou, and Lichao Sun

[Preprint]

Watermarking Classification Dataset for Copyright Protection
Yixin Liu*, Hongsheng Hu*, Xuyun Zhang, Lichao Sun

[Preprint]

BadGPT: Exploring Security Vulnerabilities of ChatGPT via Backdoor Attacks to InstructGPT
Jiawen Shi, Yixin Liu, Pan Zhou and Lichao Sun

[NDSS 2023 Poster]

Securing Biomedical Images from Unauthorized Training with Anti-Learning Perturbation
Yixin Liu, Haohui Ye, Lichao Sun

[NDSS 2023 Poster]

SEAT: Stable and Explainable Attention
Lijie Hu*, Yixin Liu*, Ninghao Liu, Mengdi Huai, Lichao Sun and Di Wang

[Paper] AAAI 2023 Oral

Conditional Automated Channel Pruning for Deep Neural Networks
Yixin Liu, Yong Guo, Jiaxin Guo, Luoqian Jiang, Jian Chen

[IEEE Signal Processing Letters]

Meta-Pruning with Reinforcement Learning
Yixin Liu; Advisor: Jian Chen

[Bachelor Thesis]

Priority Prediction of Sighting Report Using Machine Learning Methods
Yixin Liu, Jiaxin Guo, Jieyang Dong, Luoqian Jiang, Haoyuan Ouyang; Advisor: Han Huang

[IEEE SEAI 2021; Finalist Award in MCM/ICM 2021]