Tzu-Han Lin
MS Student @ National Taiwan University
Visiting Research Intern @ University of Virginia
Seeking PhD positions for Fall 2026!
I am an M.S. student at the Graduate Institute of Networking and Multimedia, National Taiwan University, fortunate to be advised by Prof. Yun-Nung (Vivian) Chen. I am also currently a visiting research intern at the University of Virginia, advised by Prof. Yu Meng.
I received my bachelor’s degree from National Taiwan University, during which I worked with Prof. Hung-yi Lee and Prof. Yun-Nung (Vivian) Chen.
Research Focus: My research centers on natural language processing, particularly Large Language Models (LLMs). I am interested in developing LLMs and agents that are resource-efficient and reliable.
Nowadays, I think about:
- Language Agents:
- Trustworthiness: Beyond task completion, how can we make the behavior of agents more interpretable and faithful to users?
- Efficiency: How can we control or reduce unnecessary search calls in long-horizon search agents (e.g., DeepResearch agents)? How should we quantify and balance the trade-off between reasoning and search?
- Self-Improving/Self-Evolving AI Systems:
- Conditions for Autonomous Improvement: What are the necessary conditions for an AI system to reliably and sustainably self-improve?
- Alignment: How can we ensure that self-improving systems remain aligned with human values as they surpass the level at which humans can easily provide direct supervision?
- Verification and Evaluation at Scale: How can we efficiently verify and evaluate progress on long-horizon, open-ended tasks that are inherently difficult to assess (e.g., scientific research agents)?
Misc: When I am not doing research, I like to listen to different genre of musics, especially Blues, Rock, and Grunge. I’m a huge fan of John Mayer and I love to play eletric guitars!
news
| Dec 18, 2025 | Our new preprint AdaSearch: Balancing Parametric Knowledge and Search in Large Language Models via Reinforcement Learning is out! We investigate how LLM search agents should balance parametric knowledge and search, and propose AdaSearch, a framework that teaches agents to explicitly decide when to search via RL. It achieves superior self-knowledge awareness without complex reward engineering, and provides transparent decision-making rationales. |
|---|---|
| May 27, 2025 | Excited to share that I have joined Prof. Yu Meng’s group as a visiting research intern, and I’m grateful for the opportunity to work with the team! |
| Feb 19, 2025 | Our new preprint Transferring Textual Preferences to Vision-Language Understanding through Model Merging is out! We show that text scalar RMs can be merged into Vision LLMs to build VL-RMs. (Update 05/2025: The paper is accepted to ACL 2025 Main.) |
| Jul 01, 2024 | Our preprint DogeRM: Equipping Reward Models with Domain Knowledge through Model Merging is out! We show that scalar reward models can be merged with intruction-tuned LLMs to derive domain-specific reward models w/o training! (Update 09/2024: The paper is accepted to EMNLP 2024 Main.) |
| Jan 04, 2024 | Our preprint PEFT for Speech: Unveiling Optimal Placement, Merging Strategies, and Ensemble Techniques is out! (Update 02/2024: The paper is accepted to ICASSP 2024 SASB Workshop.) |