Yichen (Zach) Wang   王奕辰

yichenzw at uchicago dot edu



I am a second-year CS PhD student at the University of Chicago, advised by Prof. Mina Lee in the UChicago C&I group. My current research focuses on improving generative diversity and empowering exploration and creativity in large language models for open-ended tasks, including writing, thinking, and human collaboration.

I serves as the student organizer of UChicago/TTIC NLP Seminar. Previously, I worked with Prof. Tianxing He and Prof. Yulia Tsvetkov as an intern at UW NLP. I completed my undergraduate degree at Xi’an Jiaotong University CS Honors Program, where I worked with Prof. Xiaoming Liu, Prof. Chao Shen, and Prof. Minnan Luo, and co-led the LUD research group. I have also been a visiting student at UC Berkeley, researching with Dr. Kevin Yang and Prof. Dan Klein.

I am actively seeking summer 2026 research internships on LLM.


news

Nov. 6, 2025: 🔎 I am actively seeking summer 2026 research internships on LLM. I would appreciate if you know any opportunities!
Nov. 5, 2025: Our paper on analysing misinformation in reasoning is presenting at EMNLP 2025 this week!
July 25, 2025: Our research on jailbreaking vision-language models will be on oral presentation at ACL 2025!
Oct. 1, 2024: 🌅 Starting the PhD journey at UChicago! Can't wait to learn a loooooot!
May 18, 2024: ACL 2024 accepts three of our papers (2 main 1 findings)! Please feel free to check them out! And we also won a competition at the SDP workshop. See you in Bangkok this summer!
Apr. 16, 2024: 💕 Here marks the end of my application season. My sincerest thanks to all who helped and supported me -- families, friends, advisors, mentors, faculties (especially those I applied to), and mates. I'm super excited for my new journey!
Mar. 24, 2024: SemStamp, the semantic watermark, is now accepted by NAACL24! Fantastic work by Abe, and he is seeking a PhD chance in Fall 25. Please consider him!
Dec. 11, 2023: Happy to share that AAAI24 accepts our paper DP2O on prompt optimization!
Oct. 10, 2023: Two papers accepted by EMNLP23! All applause and thanks to my co-authors! And I'll be in Singapore on Dec.!
Sep. 19, 2023: Today is the date of birth of my academic webpage! Working towards the application season!


publications

  • Optimizing Diversity and Quality through Base-Aligned Model Collaboration
    Yichen Wang=, Chenghao Yang=, Tenghao Huang=, Muhao Chen, Jonathan May, Mina Lee
    arxiv:2511
    Alignment improves large language models (LLMs)' output quality but at the cost of diversity. We propose that collaboration between base and aligned models can achieve both diversity and quality. We introduce BACo (Base-Aligned Model Collaboration), an inference-time framework that employs token-level routing strategies based on prediction uncertainty and semantic role, achieving improvements in a single pass without retraining or multi-sampling. Experiments across three open-ended generation tasks and 13 metrics show BACo consistently surpasses state-of-the-art baselines, achieving a 21.3% joint improvement in diversity and quality, confirmed by human evaluations.
    Citation // Website // Code // Data //
    Reading List: Awesome LLM Diversity -- A curated list of papers and resources on LLM diversity, to cover literature from various perspectives including linguistic, value pluralism, exploration in RL, and human-LLM interaction, etc.
  • Unraveling Misinformation Propagation in LLM Reasoning
    Yiyang Feng=, Yichen Wang=, Shaobo Cui, Boi Faltings, Mina Lee, Jiawei Zhou
    EMNLP 2025 Findings
    We investigate how misinformation from user inputs, which is prevalent in real-world interactions, propagates through LLMs' reasoning processes, focusing on math reasoning as a case study. We analyze misinformation's impact on intermediate steps and final answers, and examine LLMs' ability on correction. Results show LLMs correct misinformation less than half the time despite possessing correct internal knowledge and explicit instruction, causing accuracy drops of 10.02%-72.20%. We explore mitigation methods and suggest that fine-tuning on synthetic early-stage, factual correction data can effectively mitigate misinformation propagation.
    Citation // Website
  • Jailbreak Large Vision-Language Models Through Multi-Modal Linkage
    Yu Wang, Xiaofei Zhou, Yichen Wang, Geyuan Zhang, Tianxing He
    ACL 2025 Oral
    State-of-the-art VLMs like GPT-4o can be jailbroken at the linkage between multiple modalities. We propose Multi-Modal Linkage (MML) Attack that uses an encryption-decryption process across text and image modalities to hide malicious content and frames within benign scenarios like video game production. Experiments demonstrate attack success rates of 97.80% on SafeBench, 98.81% on MM-SafeBench, and 99.07% on HADES-Dataset against GPT-4o.
    Citation
  • Can A Society of Generative Agents Simulate Human Behavior and Inform Public Health Policy? A Case Study on Vaccine Hesitancy
    Abe Bohan Hou, Hongru Du, Yichen Wang, Jingyu Zhang, Zixiao Wang, Paul Pu Liang, Daniel Khashabi, Lauren Gardner, Tianxing He
    COLM 2025
    We explore how to apply a sandbox society with generative agents to model human behavior for assessing public policies. We introduce VacSim, a framework that uses 100 LLM agents to simulate health-related decision-making, with vaccine hesitancy as a case study. We instantiate agents with demographics, connect them via a social network, and evaluate public health interventions. Experiments indicate that LLMs can simulate aspects of human behavior but face real-world alignment challenges such as demographic inconsistencies, highlighting both the potential and limitations of LLM-driven social simulation for policy development.
    Citation
  • HACo-Det: A Study Towards Fine-Grained Machine-Generated Text Detection under Human-AI Coauthoring
    Zhixiong Su=, Yichen Wang=, Herun Wan, Zhaohan Zhang, Minnan Luo
    ACL 2025
    We explore the possibility of fine-grained machine-generated text detection under human-AI coauthoring. We adapt existing document-level detectors to fine-grained detection and evaluate them on the word-level HACo-Det dataset we built. The results show that metric-based methods significantly underperform, and all methods face challenges in detecting coauthored texts.
    Citation
  • Concentrate Attention: Towards Domain-Generalizable Prompt Optimization for Language Models
    Chengzhengxu Li, Xiaoming Liu, Zhaohan Zhang, Yichen Wang, Chen Liu, Yu Lan, and Chao Shen
    NIPS 2024
    We conduct a pilot study on prompt optimization generalization and find two co-relation rules with LM's attention weight distributions. We then offer a new objective, concentration, representing the strength and stability of lookback attention to the prompt. Adapting it to popular soft and hard prompt optimization methods shows good improvement.
    Citation
  • Stumbling Blocks: Stress Testing the Robustness of Machine-Generated Text Detectors Under Attacks
    Yichen Wang, Shangbin Feng, Abe Bohan Hou, Xiao Pu, Chao Shen, Xiaoming Liu, Yulia Tsvetkov, and Tianxing He
    ACL 2024   🌟 best paper AC nomination 🌟 meta score = 5/5
    We comprehensively study the robustness of popular machine-generated text detectors under attacks from diverse categories: editing, paraphrasing, prompting, and co-generating. Our experiments reveal that all detectors exhibit different loopholes. Further, we investigate the reasons behind these defects and propose initial out-of-the-box patches.
    Citation // Code // Dataset // Poster
  • k-SemStamp: A Clustering-Based Semantic Watermark for Detection of Machine-Generated Text
    Abe Bohan Hou, Jingyu Zhang, Yichen Wang, Daniel Khashabi, and Tianxing He
    ACL 2024 Findings
    We propose k-SemStamp, a simple yet effective enhancement of SemStamp, utilizing k-means clustering as an alternative of LSH to partition the embedding space with awareness of inherent semantic structure.
    Citation
  • Does DetectGPT Fully Utilize Perturbation? Bridging Selective Perturbation to Fine-tuned Contrastive Learning Detector would be Better
    Shengchao Liu, Xiaoming Liu, Yichen Wang, Zehua Cheng, Chengzhengxu Li, Zhaohan Zhang, Yu Lan, and Chao Shen
    ACL 2024
    We propose a novel fine-tuned machine-generated text detector, Pecola, bridging metric-based and fine-tuned methods by contrastive learning on selective perturbation further than DetectGPT.
    Citation
  • SemStamp: A Semantic Watermark with Paraphrastic Robustness for Text Generation
    Abe Bohan Hou, Jingyu Zhang, Tianxing He, Yichen Wang, Yung-Sung Chuang, Hongwei Wang, Lingfeng Shen, Benjamin Van Durme, Daniel Khashabi, and Yulia Tsvetkov
    NAACL 2024
    Existing watermarking algorithms are vulnerable to paraphrase attacks because of their token-level design. To address this issue, we propose SemStamp, a robust sentence-level semantic watermarking algorithm based on locality-sensitive hashing (LSH), which partitions the semantic space of sentences.
    Citation
  • Dialogue for Prompting: a Policy-Gradient-Based Discrete Prompt Optimization for Few-shot Learning
    Chengzhengxu Li, Xiaoming Liu, Yichen Wang, Duyi Li, Yu Lan, and Chao Shen
    AAAI 2024
    We propose a dialogue-comprised policy-gradient-based discrete prompt optimization (DP2O) method with dialogue prompt alignment and reinforcement learning to efficiently and effectively generate prompt demonstrations.
    Citation
  • Improving Pacing in Long-Form Story Planning
    Yichen Wang, Kevin Yang, Xiaoming Liu, and Dan Klein
    EMNLP 2023 Findings
    Existing LLM-based systems for writing long-form stories or story outlines frequently suffer from unnatural pacing, resulting in a jarring experience for the reader. We propose a Concrete Outline Control (CONCOCT) system to improve pacing when automatically generating story outlines. Compared to a baseline hierarchical outline generator, humans judge CONCOCT’s pacing to be more consistent over 57% of the time across multiple outline lengths, and the gains also translate to downstream stories.
    Citation // Code // Dataset // Poster
  • CoCo: Coherence-Enhanced Machine-Generated Text Detection Under Data Limitation With Contrastive Learning
    Xiaoming Liu=, Zhaohan Zhang=, Yichen Wang=, Hang Pu, Yu Lan, and Chao Shen
    EMNLP 2023
    We present a coherence-based contrastive learning model named CoCo to detect the possible machine-generated texts (MGTs) under the low-resource scenario. We encode coherence information in the form of graph into the text representation and employ an improved contrastive learning framework. Our approach outperforms the state-of-the-art methods at least 1.23%. Also, we surprisingly find that MGTs originated from up-to-date language models could be easier to detect than these from previous models, in our experiments, and we propose some preliminary explanations.
    Citation // Code // Dataset // Poster

competition