Fanqi Wan (万凡琦)

Hello! My name is Fanqi Wan, and I am a final-year MS student at Sun Yat-sen University, advised by Prof. Xiaojun Quan. Before this, I received my Bachelor's degree from Xi'an Jiaotong University. Previously, I conducted my internship at Tencent AI Lab (2023.03-2024.05), where I am mentored by Dr. Xinting Huang and Dr. Wei Bi. Now, I am currently conducting my internship at ByteDance Doubao (Seed) LLM Team (2024.06-Now).

Email / CV(en/zh) / Google Scholar / Semantic Scholar / GitHub / HF / X

profile photo
Research

My main research interests focused on deep learning for natural language generation. Previously, my work primarily focused on dialogue systems. After the emergence of large language models (LLMs), my research direction shifted towards efficient alignment (e.g., developing LLMs for specific domains, mitigating hallucinations of LLMs, self-improving of LLMs) and knowledge fusion (e.g., combining the strengths of LLMs with diverse structures/scales/functionarities). Representative papers are highlighted.

Knowledge Fusion
PontTuset FuseChat: Knowledge Fusion of Chat Models
Fanqi Wan, Longguang Zhong, Ziyi Yang, Ruijun Chen, Xiaojun Quan
Tech Report, 2024
[GitHub] / [HF] / [Paper] / [Featured by AK] / [HF Daily Papers] / [机器之心]

We propose FuseChat, an extended framework of FuseLLM to integrate the collective knowledge and individual strengths of multiple structure- and scale-varied chat LLMs into a more powerful chat LLM. FuseChat-7B is comparable to the larger Mixtral-8x7B-Instruct and and approaches GPT-3.5-Turbo-1106 on MT-Bench.

PontTuset ProFuser: Progressive Fusion of Large Language Models
Tianyuan Shi, Fanqi Wan, Canbin Huang, Xiaojun Quan, Chenliang Li, Ming Yan, Ji Zhang
Tech Report, 2024
[GitHub] / [HF] / [Paper]

We introduce a novel approach that enhances the fusion process by incorporating both the training and inference modes from multiple structurally diverse LLMs.

PontTuset Knowledge Fusion of Large Language Models
Fanqi Wan, Xinting Huang, Deng Cai, Xiaojun Quan, Wei Bi, Shuming Shi
ICLR, 2024
[GitHub] / [HF] / [Paper] / [Featured by elvis] / [Featured by AIDB] / [机器之心]

We propose FuseLLM to create a unified model that combines the distinctive strengths of multiple structurally diverse LLMs. FuseLLM-7B surpasses Llama-2-7B on 12 benchmarks, including commonsense, reasoning, question-answering, and code generation.

Efficient Alignment
PontTuset Self-Evolution Fine-Tuning for Policy Optimization
Ruijun Chen, Jiehao Liang, Shiping Gao, Fanqi Wan, Xiaojun Quan
EMNLP Findings, 2024
[GitHub] / [Paper]

We introduce self-evolution fine-tuning for policy optimization, which eliminates the need for annotated data samples during alignment.

PontTuset Knowledge Verification to Nip Hallucination in the Bud
Fanqi Wan, Xinting Huang, Leyang Cui, Xiaojun Quan, Wei Bi, Shuming Shi
EMNLP, 2024
[GitHub] / [HF] / [Paper]

We introduce Knowledge Consistent Alignment to verify and minimize the knowledge inconsistency between external knowledge in the alignment data and the intrinsic knowledge embedded in foundation LLMs, thus mitigating hallucinations before alignment.

PontTuset Explore-Instruct: Enhancing Domain-Specific Instruction Coverage through Active Exploration
Fanqi Wan, Xinting Huang, Tao Yang, Xiaojun Quan, Wei Bi, Shuming Shi
EMNLP, 2023
[GitHub] / [HF] / [Paper] / [Paper Weekly]

We propose a novel approach to enhance the domain-specific instruction coverage by utilizing LLMs to explore the domain space from both breadth and depth automatically. Explore-Instruct outperforms Self-Instruct in three specific domains.

Dialogue Systems
PontTuset Retrieval-Generation Alignment for End-to-End Task-Oriented Dialogue System
Weizhou Shen, Yingqi Gao, Canbin Huang, Fanqi Wan, Xiaojun Quan, Wei Bi
EMNLP, 2023
[GitHub] / [Paper]

We introduce maximal marginal likelihood for retriever training to address the retrieval-generation misalignment in end-to-end task-oriented dialogue systems.

PontTuset Multi-Grained Knowledge Retrieval for End-to-End Task-Oriented Dialog
Fanqi Wan, Weizhou Shen, Ke Yang, Xiaojun Quan, Wei Bi
ACL, 2023
[GitHub] / [HF] / [Paper]

We propose a multi-grained knowledge retriever and introduce a novel distillation objective for retriever training. MAKER achieves SOTA performance on MultiWOZ 2.1 and CamRest with both condensed KB and full KB.

Misc.
PontTuset BlockPruner: Fine-grained Pruning for Large Language Models
Longguang Zhong, Fanqi Wan, Ruijun Chen, Xiaojun Quan, Liangzhi Li
Tech Report, 2024
[GitHub] / [Paper]

We propose a training-free structured pruning approach for fine-grained LLMs pruning.

PontTuset PsyCoT: Psychological Questionnaire as Powerful Chain-of-Thought for Personality Detection
Tao Yang, Tianyuan Shi, Fanqi Wan, Xiaojun Quan, Qifan Wang, Bingzhe Wu, Jiaxiang Wu
EMNLP Findings, 2023
[GitHub] / [Paper]

We propose a novel method for zero-shot personality detection in a multi-turn dialogue manner.

PontTuset Clustering-Aware Negative Sampling for Unsupervised Sentence Representation
Jinghao Deng, Fanqi Wan, Tao Yang, Xiaojun Quan, Rui Wang
ACL Findings, 2023
[GitHub] / [Paper]

We propose a novel method that incorporates cluster information into contrastive learning for unsupervised sentence representation learning.

Education

MS Student in Computer Science, Sun Yat-sen University (2022.9-now).

Bachelor of Automation, Xi'an Jiaotong University (2018.9-2022.6).

Experience

Research Intern at LLM Team, ByteDance Doubao (Seed). (2024.6-now).

Research Intern at NLPC, Tencent AI Lab, supervised by Dr. Xinting Huang and Dr. Wei Bi. (2023.3-2024.5).

Commercial Projects on E-Commerce Platforms, Vipshop, supervised by Dr. Rui Wang. (2022.4-2023.1).

Academic Competitions

2nd Prize on 2023 Xingzhi Cup Deep Learning Model Interpretability Task. (Team Leader) [Task]

2nd Prize on 2022 IFLYTEK AI Developer Competition Paper Abstracts Classification Task. (Team Leader) [Task]

3nd Prize on 2022 Ali Lingjie E-commerce Search Algorithm Competition. (Team Leader) [Task]

Academic Services

NeurIPS Reviewer, 2024-.

ICLR Reviewer, 2025-.

AISTATS Reviewer, 2025-.

Selected Awards

National Scholarship, Sun Yat-sen University, 2023.

Outstanding Award for Tencent AI Lab Rhino-Bird Focused Research Program, Tencent, 2023.

Excellent Graduate Students, Xi'an Jiaotong University, 2022.

National Scholarship, Xi'an Jiaotong University, 2018.


Website's code is from Jon Barron.