Tzu-Han Lin

prof_pic.jpg

MS Student @ National Taiwan University
Visiting Research Intern @ University of Virginia
Seeking PhD positions for Fall 2026!

I am an M.S. student at the Graduate Institute of Networking and Multimedia, National Taiwan University, fortunate to be advised by Prof. Yun-Nung (Vivian) Chen. I am also currently a visiting research intern at the University of Virginia, advised by Prof. Yu Meng.

I received my bachelor’s degree from National Taiwan University, during which I worked with Prof. Hung-yi Lee and Prof. Yun-Nung (Vivian) Chen.

Research Focus: My research centers on natural language processing, particularly Large Language Models (LLMs). I am interested in developing LLMs and agents that are resource-efficient and reliable.

Nowadays, I think about:

  1. Language Agents:
    • Trustworthiness: Beyond task completion, how can we make the behavior of agents more interpretable and faithful to users?
    • Efficiency: How can we control or reduce unnecessary search calls in long-horizon search agents (e.g., DeepResearch agents)? How should we quantify and balance the trade-off between reasoning and search?
  2. Self-Improving/Self-Evolving AI Systems:
    • Conditions for Autonomous Improvement: What are the necessary conditions for an AI system to reliably and sustainably self-improve?
    • Alignment: How can we ensure that self-improving systems remain aligned with human values as they surpass the level at which humans can easily provide direct supervision?
    • Verification and Evaluation at Scale: How can we efficiently verify and evaluate progress on long-horizon, open-ended tasks that are inherently difficult to assess (e.g., scientific research agents)?

Misc: When I am not doing research, I like to listen to different genre of musics, especially Blues, Rock, and Grunge. I’m a huge fan of John Mayer and I love to play eletric guitars!

news

Dec 18, 2025 Our new preprint AdaSearch: Balancing Parametric Knowledge and Search in Large Language Models via Reinforcement Learning is out! We investigate how LLM search agents should balance parametric knowledge and search, and propose AdaSearch, a framework that teaches agents to explicitly decide when to search via RL. It achieves superior self-knowledge awareness without complex reward engineering, and provides transparent decision-making rationales.
May 27, 2025 Excited to share that I have joined Prof. Yu Meng’s group as a visiting research intern, and I’m grateful for the opportunity to work with the team!
Feb 19, 2025 Our new preprint Transferring Textual Preferences to Vision-Language Understanding through Model Merging is out! We show that text scalar RMs can be merged into Vision LLMs to build VL-RMs. (Update 05/2025: The paper is accepted to ACL 2025 Main.)
Jul 01, 2024 Our preprint DogeRM: Equipping Reward Models with Domain Knowledge through Model Merging is out! We show that scalar reward models can be merged with intruction-tuned LLMs to derive domain-specific reward models w/o training! (Update 09/2024: The paper is accepted to EMNLP 2024 Main.)
Jan 04, 2024 Our preprint PEFT for Speech: Unveiling Optimal Placement, Merging Strategies, and Ensemble Techniques is out! (Update 02/2024: The paper is accepted to ICASSP 2024 SASB Workshop.)

selected publications

  1. AdaSearch: Balancing Parametric Knowledge and Search in Large Language Models via Reinforcement Learning
    Tzu-Han Lin, Wei-Lin Chen, Chen-An Li, Hung-yi Lee, Yun-Nung Chen, and Yu Meng
    arXiv preprint arXiv:2512.16883, 2025
  2. Transferring Textual Preferences to Vision-language Understanding Through Model Merging
    Chen-An Li, Tzu-Han Lin, Yun-Nung Chen, and Hung-yi Lee
    In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2025
  3. DogeRM: Equipping Reward Models with Domain Knowledge through Model Merging
    Tzu-Han Lin, Chen-An Li, Hung-Yi Lee, and Yun-Nung Chen
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024
  4. PEFT for Speech: Unveiling Optimal Placement, Merging Strategies, and Ensemble Techniques
    Tzu-Han Lin*, How-Shing Wang*, Hao-Yung Weng**, Kuang-Chen Peng**, Zih-Ching Chen, and Hung-yi Lee
    In 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW), 2024