Yixin Liu (刘奕鑫)

4th-year CSE Ph.D. student at Lehigh University (Advisor: Prof. Lichao Sun)
Research: Trustworthy ML, Generative AI Safety, Content Protection/Provenance
Industry: Dolby Labs (8 months), Samsung Research America (11 months)
B.E. Software Engineering, South China University of Technology (2022)

News & Highlights
  • [Seeking Position] I am actively looking for full-time research scientist/engineer positions starting May 2026. My expertise spans LLMs, reinforcement fine-tuning (Group Relative Policy Optimization/GRPO), synthetic content detection, audio language models, AI safety, content protection, watermarking, and trustworthy ML. Please reach out if you have opportunities!
  • [2025.08] Developed explainable audio synthetic content detection system at Dolby Labs using LLMs with reinforcement learning, achieving better explainability, generalizability, and effectiveness.
  • [2025.01] Our XAttnMark is accepted by ICML'25! State-of-the-art neural audio watermarking achieving joint detection and attribution. [Virtual Poster]
  • [2024.09] Started research internship at Dolby Labs, working on audio watermarking and content protection for Universal Music Group.
  • [2024.07] Invited Talk at Microsoft ASG Research Day on "Adversarial Perturbation in Personalized Diffusion Models".
  • [2024.06] Presented MetaCloak as CVPR'24 Oral - watch the talk!
  • [2024.05] Our FViT is accepted by ICML'24 as Spotlight!
  • [2024.05] Completed internship at Samsung Research America on Graph-based RAG for log analysis.
  • [2023.12] Our Stable Unlearnable Example (SEM) is accepted by AAAI'24! Achieving 3.91× speedup with improved protection efficacy.
Research Interest

I am broadly interested in content protection in the era of generative AI and trustworthy ML. My research focuses on the intersection between powerful generative AI and copyright protection, developing data-centric approaches to safeguard user data from unauthorized exploitation and provide robust source verification.

  • Proactive UGC Protection [MetaCloak (CVPR'24 Oral), DiffShortcut, SEM (AAAI'24), EditShield (ECCV'24), MUE (ICML'24 Workshop), GraphCloak, Linear Solver Analysis]: User-generated content faces unprecedented threats from unauthorized AI training. By exploiting the fundamental vulnerability that neural networks are not robust to small input perturbations, we develop protective mechanisms that prevent unauthorized model training while preserving data utility for legitimate use.
  • AIGC Watermarking and Detection [XAttnMark (ICML'25), TextMarker]: As generative models blur the boundary between real and fake content, robust authentication becomes critical. We develop watermarking techniques that enable AI content attribution and source verification across modalities, including state-of-the-art neural audio watermarking with joint detection and attribution capabilities.
  • Robust Explainable AI [SEAT (AAAI'23 Oral), FViTs (ICML'24 Spotlight)]: The exploration of adversarial learning and explainability mechanisms enhances our understanding of model vulnerabilities and interpretability. SEAT addresses attention mechanism instability, providing stable and explainable attention for NLP tasks. FViTs develops faithful vision transformers through denoised diffusion smoothing, ensuring robust attention maps under adversarial attacks. These works advance our understanding of adversarial learning, robustness red-teaming, and robust explainable AI foundations.
Professional Experience
  • Dolby Labs - Research Intern (Sep 2024 - Apr 2025, May 2025 - Aug 2025)
    Working on robust audio watermarking for content protection with Universal Music Group. Developed XAttnMark achieving state-of-the-art detection and attribution performance. Extended work to explainable audio synthetic content detection using LLMs with reinforcement learning (using VERL), focusing on watermark-free scenarios.
  • Samsung Research America - Research Intern (May 2024 - Aug 2024)
    Developed graph-based RAG system for log analysis, achieving +16 comprehensiveness score improvement. Also worked on DiffShortcut for defending protective perturbations in diffusion models.
  • Samsung Research America - Research Intern (May 2023 - Nov 2023)
    Proposed efficient defensive perturbation generation methods for data protection against diffusion models, resulting in MetaCloak (CVPR'24 Oral) and GraphCloak for graph data protection.
  • Lehigh University - Teaching Assistant
    CSE 017 Java Programming (Spring 2023), CSE 007 Python Programming (Spring 2024)
Invited Talks & Presentations
  • ICML 2025 Poster - "XAttnMark: Learning Robust Audio Watermarking with Cross-Attention" [Virtual Poster]
  • Dolby Lab Tech Summit - "Robust Audio Watermarking for the Music Industry" (June 2025) [Slides]
  • Microsoft ASG Research Day - "Adversarial Perturbation in Personalized Diffusion" (invited by Dr. Tianyi Chen, July 2024) [Slides]
  • CVPR 2024 Oral - "MetaCloak: Preventing Unauthorized T2I Diffusion Synthesis" (June 2024) [Video] [Slides]
Reviewer Service

NeurIPS'23'24, KDD'23'25, CVPR'24'25, ICML'24'25, ECCV'24 (Outstanding Reviewer), ICLR'25, ICASSP'25, IEEE TIP

Publications ( show selected / show all by topic / show all by date )

Topics: Unauthorized Exploitation / NLP Safety / Explainable AI / Model Compresssion / Applications (*/†: indicates equal contribution.)

XAttnMark: Learning Robust Audio Watermarking with Cross-Attention
Yixin Liu, Lie Lu, Jihui Jin, Lichao Sun, Andrea Fanelli

[Project Page] [Paper] [ICML Talk] ICML 2025

MetaCloak: Preventing Unauthorized Subject-driven Text-to-image Diffusion-based Synthesis via Meta-learning
Yixin Liu, Chenrui Fan, Yutong Dai, Xun Chen, Pan Zhou, Lichao Sun

[CVPR 2024 Oral]

Medical Unlearnable Examples: Securing Medical Data from Unauthorized Traning via Sparsity-Aware Local Masking
Weixiang Sun, Yixin Liu, Zhiling Yan, Kaidi Xu, Lichao Sun

[ICML'24 Next Gen AI Safety 2024 Workshop]

Stable Unlearnable Example: Enhancing the Robustness of Unlearnable Examples via Stable Error-Minimizing Noise
Yixin Liu, Kaidi Xu, Xun Chen, Lichao Sun

[AAAI 2024]

Improving Faithfulness for Vision Transformers
Lijie Hu*, Yixin Liu*, Ninghao Liu, Mengdi Huai, Lichao Sun and Di Wang

[ICML 2024 Spotlight]

GraphCloak: Safeguarding Graph-structured Data from Unauthorized Exploitation
Yixin Liu, Chenrui Fan, Xun Chen, Pan Zhou, and Lichao Sun

[Preprint]

Watermarking Classification Dataset for Copyright Protection
Yixin Liu*, Hongsheng Hu*, Xuyun Zhang, Lichao Sun

[Preprint]

BadGPT: Exploring Security Vulnerabilities of ChatGPT via Backdoor Attacks to InstructGPT
Jiawen Shi, Yixin Liu, Pan Zhou and Lichao Sun

[NDSS 2023 Poster]

Securing Biomedical Images from Unauthorized Training with Anti-Learning Perturbation
Yixin Liu, Haohui Ye, Lichao Sun

[NDSS 2023 Poster]

SEAT: Stable and Explainable Attention
Lijie Hu*, Yixin Liu*, Ninghao Liu, Mengdi Huai, Lichao Sun and Di Wang

[Paper] AAAI 2023 Oral

Conditional Automated Channel Pruning for Deep Neural Networks
Yixin Liu, Yong Guo, Jiaxin Guo, Luoqian Jiang, Jian Chen

[IEEE Signal Processing Letters]

Meta-Pruning with Reinforcement Learning
Yixin Liu; Advisor: Jian Chen

[Bachelor Thesis]

Priority Prediction of Sighting Report Using Machine Learning Methods
Yixin Liu, Jiaxin Guo, Jieyang Dong, Luoqian Jiang, Haoyuan Ouyang; Advisor: Han Huang

[IEEE SEAI 2021; Finalist Award in MCM/ICM 2021]