Fanqi Wan (万凡琦)

Hello! My name is Fanqi Wan, and I am a second-year MS student at Sun Yat-sen University, advised by Prof. Xiaojun Quan. Before this, I received my Bachelor's degree from Xi'an Jiaotong University. I am currently conducting my internship at Tencent AI Lab, where I am mentored by Dr. Xinting Huang and Dr. Wei Bi.

Email / CV(en/zh) / Google Scholar / GitHub / HF / LinkedIn / Twitter

profile photo
Research

My main research interests focused on deep learning for natural language generation. Previously, my work primarily focused on dialogue systems. After the emergence of large language models (LLMs), my research direction shifted towards instruction-tuning (e.g., developing LLMs for specific domains, mitigating hallucinations of LLMs) and model fusion (e.g., combining the capabilities of multiple structurally diverse LLMs). Representative papers are highlighted.

Model Fusion
PontTuset FuseChat: Knowledge Fusion of Chat Models
Fanqi Wan, Ziyi Yang, Longguang Zhong, Xiaojun Quan, Xinting Huang, Wei Bi,
Tech Report, 2024
[GitHub] / [HF] / [Paper] / [Featured by AK] / [HF Daily Papers] / [机器之心]

We propose FuseChat, an extended framework of FuseLLM to integrate the collective knowledge and individual strengths of multiple structure- and scale-varied chat LLMs into a more powerful chat LLM. FuseChat-7B achieves 8.22 on MT-Bench, which is the current SOTA 7B LLM.

PontTuset Knowledge Fusion of Large Language Models
Fanqi Wan, Xinting Huang, Deng Cai, Xiaojun Quan, Wei Bi, Shuming Shi
ICLR, 2024
[GitHub] / [HF] / [Paper] / [Featured by elvis] / [Featured by AIDB] / [机器之心]

We propose FuseLLM to create a unified model that combines the distinctive strengths of multiple structurally diverse LLMs. FuseLLM-7B surpasses Llama-2-7B on 12 benchmarks, including commonsense, reasoning, question-answering, and code generation.

Instruction-Tuning
PontTuset Knowledge Verification to Nip Hallucination in the Bud
Fanqi Wan, Xinting Huang, Leyang Cui, Xiaojun Quan, Wei Bi, Shuming Shi
Tech Report, 2024
[GitHub] / [HF] / [Paper]

We introduce Knowledge Consistent Alignment to verify and minimize the knowledge inconsistency between external knowledge in the alignment data and the intrinsic knowledge embedded in foundation LLMs, thus mitigating hallucinations before alignment.

PontTuset Explore-Instruct: Enhancing Domain-Specific Instruction Coverage through Active Exploration
Fanqi Wan, Xinting Huang, Tao Yang, Xiaojun Quan, Wei Bi, Shuming Shi
EMNLP, 2023
[GitHub] / [HF] / [Paper] / [Paper Weekly]

We propose a novel approach to enhance the domain-specific instruction coverage by utilizing LLMs to explore the domain space from both breadth and depth automatically. Explore-Instruct outperforms Self-Instruct in three specific domains.

Dialogue Systems
PontTuset Retrieval-Generation Alignment for End-to-End Task-Oriented Dialogue System
Weizhou Shen, Yingqi Gao, Canbin Huang, Fanqi Wan, Xiaojun Quan, Wei Bi
EMNLP, 2023
[GitHub] / [Paper]

We introduce maximal marginal likelihood for retriever training to address the retrieval-generation misalignment in end-to-end task-oriented dialogue systems.

PontTuset Multi-Grained Knowledge Retrieval for End-to-End Task-Oriented Dialog
Fanqi Wan, Weizhou Shen, Ke Yang, Xiaojun Quan, Wei Bi
ACL, 2023
[GitHub] / [HF] / [Paper]

We propose a multi-grained knowledge retriever and introduce a novel distillation objective for retriever training. MAKER achieves SOTA performance on MultiWOZ 2.1 and CamRest with both condensed KB and full KB.

Misc.
PontTuset PsyCoT: Psychological Questionnaire as Powerful Chain-of-Thought for Personality Detection
Tao Yang, Tianyuan Shi, Fanqi Wan, Xiaojun Quan, Qifan Wang, Bingzhe Wu, Jiaxiang Wu
EMNLP Findings, 2023
[GitHub] / [Paper]

We propose a novel method for zero-shot personality detection in a multi-turn dialogue manner.

PontTuset Clustering-Aware Negative Sampling for Unsupervised Sentence Representation
Jinghao Deng, Fanqi Wan, Tao Yang, Xiaojun Quan, Rui Wang
ACL Findings, 2023
[GitHub] / [Paper]

We propose a novel method that incorporates cluster information into contrastive learning for unsupervised sentence representation learning.

Education

MS Student in Computer Science, Sun Yat-sen University (2022.9-now).

Bachelor of Automation, Xi'an Jiaotong University (2018.9-2022.6).

Experience

Research Intern at NLPC, Tencent AI Lab, supervised by Dr. Xinting Huang and Dr. Wei Bi. (2023.3-now).

Commercial Projects on E-Commerce Platforms, Vipshop, supervised by Dr. Rui Wang. (2022.4-2023.1).

Academic Competitions

2nd Prize on 2023 Xingzhi Cup Deep Learning Model Interpretability Task. (Team Leader) [Task]

2nd Prize on 2022 IFLYTEK AI Developer Competition Paper Abstracts Classification Task. (Team Leader) [Task]

3nd Prize on 2022 Ali Lingjie E-commerce Search Algorithm Competition. (Team Leader) [Task]

Selected Awards

Outstanding Award for Tencent AI Lab Rhino-Bird Focused Research Program, Tencent, 2023.

Excellent Graduate Students, Xi'an Jiaotong University, 2022.

National Scholarship, Xi'an Jiaotong University, 2019.


Website's code is from Jon Barron.