hero

Xiaoze Liu

ねだるな、勝ち取れ、さすれば与えられん

# About

I'm a PhD student at Purdue University, advised by Prof. Jing Gao and Prof. Xiaoqian Wang. I hold a Master's Degree with honors from Zhejiang University (China) and a Bachelor's Degree with honors from Northeastern University (China).

My research centers on the post-training and alignment paradigms of large language models (LLMs), aiming to build systems that are collaboratively capable, intrinsically trustworthy, and structurally secure. My work has been published in top-tier venues (e.g., ICLR, COLM, and ACL venues) and garnered 1,000+ citations.

I investigate two complementary frontiers:

  • (1) Collaborative Intelligence: Enhancing capabilities through collaborative learning and model merging frameworks. Key contributions include mutual reinforcement learning and dynamic ensembling, extending to federated preference optimization. I also uncover critical supply-chain vulnerabilities in model merging.
  • (2) Trustworthy Compliance: Addressing systemic risks with a focus on copyright and factuality. I engineer mechanisms for copyright compliance via reinforcement unlearning and agent-based defense, while enhancing model reliability through knowledge-based evaluation and calibration.

OPEN TO RESEARCH INTERNSHIP POSITIONS FOR SUMMER 2026! If you have any opportunities, please feel free to reach out to me via email ([myfirstname]@purdue.edu) or LinkedIn.

# News

# Recent Publications

  1. The The Trojan in the Vocabulary: Stealthy Sabotage of LLM Composition. Xiaoze Liu, Weichen Yu, Matt Fredrikson, Xiaoqian Wang, Jing Gao

  2. SUV: Scalable Large Language Model Copyright Compliance with Regularized Selective Unlearning. Tianyang Xu#, Xiaoze Liu#, Feijie Wu, Xiaoqian Wang, Jing Gao, The 2025 Conference on Language Modeling (COLM), 2025.

  3. SHIELD: Evaluation and Defense Strategies for Copyright Compliance in LLM Text Generation. Xiaoze Liu#, Ting Sun#, Tianyang Xu, Feijie Wu, Cunxiang Wang, Xiaoqian Wang, Jing Gao, The 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2024.

  4. Survey on Factuality in Large Language Models: Knowledge, Retrieval and Domain-Specificity. Cunxiang Wang#, Xiaoze Liu#, Yuanhao Yue#, Qipeng Guo, Xiangkun Hu, Xiangru Tang, Tianhang Zhang, Jiayang Cheng, Yunzhi Yao, Wenyang Gao, Xuming Hu, Zehan Qi, Yidong Wang, Linyi Yang, Jindong Wang, Xing Xie, Zheng Zhang, Yue Zhang, ACM Computing Surveys (CSUR), 2025

  5. Towards Federated RLHF with Aggregated Client Preference for LLMs. Feijie Wu, Xiaoze Liu, Haoyu Wang, Xingchen Wang, Lu Su, Jing Gao, The Thirteenth International Conference on Learning Representations (ICLR), 2025

  6. Evaluating the Factuality of Large Language Models using Large-Scale Knowledge Graphs. Xiaoze Liu, Feijie Wu, Tianyang Xu, Zhuo Chen, Yichi Zhang, Xiaoqian Wang, Jing Gao IEEE Data Engineering Bulletin, 2024

  7. SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales. EMNLP 2024 Tianyang Xu, Shujin Wu, Shizhe Diao, Xiaoze Liu, Xingyao Wang, Yangyi Chen, Jing Gao, The 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2024.

  8. CausalEval: Towards Better Causal Reasoning in Language Models. Longxuan Yu, Delin Chen, Siheng Xiong, Qingyang Wu, Qingzhen Liu, Dawei Li, Zhikai Chen, Xiaoze Liu, Liangming Pan, The 2025 Annual Conference of the Nations of the Americas Chapter of the ACL (NAACL), 2025

...

(Google Scholar)

# Education and Experience

# Academic Services

Serve as a reviewer/PC for

  • 2025: ICML, ACL Rolling Review
  • 2025: KDD, ICLR, ICML, NeurIPS, ACL Rolling Review
  • 2024: KDD, NeurIPS, ACM MM, SDM, CIKM, ISWC, ACL Rolling Review
  • 2023: KDD, NeurIPS, EMNLP

Served as a Journal reviewer for Transactions on Machine Learning Research, IEEE Transactions on Pattern Analysis and Machine Intelligence, IEEE Transactions on Knowledge and Data Engineering, IEEE Transactions on Neural Networks and Learning Systems, Pattern Recognition, Information Sciences, IEEE Transactions on Big Data