news

Dec 18, 2025 Our new preprint AdaSearch: Balancing Parametric Knowledge and Search in Large Language Models via Reinforcement Learning is out! We investigate how LLM search agents should balance parametric knowledge and search, and propose AdaSearch, a framework that teaches agents to explicitly decide when to search via RL. It achieves superior self-knowledge awareness without complex reward engineering, and provides transparent decision-making rationales.
May 27, 2025 Excited to share that I have joined Prof. Yu Meng’s group as a visiting research intern, and I’m grateful for the opportunity to work with the team!
Feb 19, 2025 Our new preprint Transferring Textual Preferences to Vision-Language Understanding through Model Merging is out! We show that text scalar RMs can be merged into Vision LLMs to build VL-RMs. (Update 05/2025: The paper is accepted to ACL 2025 Main.)
Jul 01, 2024 Our preprint DogeRM: Equipping Reward Models with Domain Knowledge through Model Merging is out! We show that scalar reward models can be merged with intruction-tuned LLMs to derive domain-specific reward models w/o training! (Update 09/2024: The paper is accepted to EMNLP 2024 Main.)
Jan 04, 2024 Our preprint PEFT for Speech: Unveiling Optimal Placement, Merging Strategies, and Ensemble Techniques is out! (Update 02/2024: The paper is accepted to ICASSP 2024 SASB Workshop.)