Yutao Mou (牟宇滔)

I am a third-year master student at PRIS Lab, Beijing University of Posts and Telecommunications, and supervised by Weiran Xu. I received the B.S. degree from Beijing University of Posts and Telecommunications, in 2021.

I have been working on building safe, reliable and scalable artificial intelligence systems. My main research interests include natural language processing and dialogue system, especially out-of-distribution (OOD) generalization, representation learning, and pre-training. Recently, I focus on large language models, such as LLM safety and alignment.

Feel free to contact me for communication and collaboration.

Email  /  Scholar  /  Github

profile photo

Publications

UEGP: Unified Expert-Guided Pre-training for Knowledge Rekindle
Yutao Mou*, Kexiang Wang, Jianhe Lin, Dehong Ma, Jun Fan, Daiting Shi, Zhicong Cheng, Gu Simiu, Dawei Yin, Weiran Xu
NAACL findings, 2024
Code / Paper

This paper first propose a new paradigm: knowledge rekindle, which aims to re-incorporate the fine-tuned expert model into the training cycle and break through the performance upper bounds of experts without introducing additional annotated data. Then we further propose a unified expert-guided pre-training (UEGP) framework for knowledge rekindle.

Decoupling Pseudo Label Disambiguation and Representation Learning for Generalized Intent Discovery
Yutao Mou*, Xiaoshuai Song*, Keqing He*, Chen Zeng, Pei Wang, Jingang Wang, Yunsen Xian, Weiran Xu
ACL, 2023
Code / Paper

This paper focuses on the generalized intent discovery task, and proposes a decoupled prototype learning framework (DPL) to decouple pseudo label disambiguation and representation learning.

Watch the Neighbors: A Unified K-Nearest Neighbor Contrastive Learning Framework for OOD Intent Discovery
Yutao Mou*, Keqing He*, Pei Wang, Yanan Wu, Jingang Wang, Wei Wu, Weiran Xu
EMNLP, 2022
Code / Paper

This paper focuses on new intent discovery and clustering task, and propose a unified K-nearest neighbor contrastive learning framework to discover OOD intents. Specifically, we design a novel K-nearest neighbor contrastive learning objective (KCL) for in-domain pre-training, and a hard negative mining strategy for self-supervised representation learning on unlabeled out-of-domain data.

UniNL: Aligning Representation Learning with Scoring Function for OOD Detection via Unified Neighborhood Learning
Yutao Mou*, Pei Wang*, Keqing He*, Yanan Wu, Jingang Wang, Wei Wu, Weiran Xu
EMNLP, 2022
Code / Paper

This paper focuses on out-of-domain (OOD) intent detection task, and propose a unified neighborhood learning framework (UniNL) to detect OOD intents. propose a unified K-nearest neighbor contrastive learning framework to discover OOD intents. Specifically, we design a K-nearest neighbor contrastive learning objective for training and introduce a KNN-based scoring function for confidence estimation. We aim to align training objective with confidence function in inference stage.

Generalized Intent Discovery: Learning from Open World Dialogue System
Yutao Mou*, Keqing He*, Yanan Wu, Pei Wang, Jingang Wang, Wei Wu, Yi Huang, Junlan Feng, Weiran Xu
COLING, 2022
Code / Paper

This paper defines a new task, Generalized Intent Discovery (GID), which aims to extend an IND intent classifier to an open-world intent set including IND and OOD intents. We hope to simultaneously classify a set of labeled IND intent classes while discovering and recognizing new unlabeled OOD types incrementally. We construct three public datasets for different application scenarios and propose two kinds of frameworks, pipeline-based and end-to-end for future work.

Disentangled Knowledge Transfer for OOD Intent Discovery with Unified Contrastive Learning
Yutao Mou*, Keqing He*, Yanan Wu*, Zhiyuan Zeng, Hong Xu, Huixing Jiang, Wei Wu, Weiran Xu
ACL, 2022
Code / Paper

Discovering Out-of-Domain (OOD) intents is essential for developing new skills in a task-oriented dialogue system. The key challenge is how to transfer prior IND knowledge to OOD clustering. This paper proposes a decoupled knowledge transfer framework for new intent discovery and clustering, which unifies the two-stage learning process of in-domain and out-of-domain data into instance discrimination and clustering discrimination tasks and bridges the gap between in-domain and out-of-domain data.

Internships

  • Sensetime Research, Beijing, China. June 2023 - December 2023.
    Research Internship, working on Large Language Model and Hallucination Correction.

  • Baidu Inc., Beijing, China. December 2022 - May 2023.
    Research Internship, focusing on search re-ranking and pre-training.

Selected Honors

  • China National Scholarship. Ministry of Education of P.R. China. 2023.

  • Excellent Graduate. Beijing University of Posts and Telecommunications. 2023.

  • Schlumberger Enterprise Scholarship. Beijing University of Posts and Telecommunications. 2022.

  • 1st Award on SereTOD Challenge 2022 track 2, EMNLP 2022

  • Excellent Graduate. Beijing University of Posts and Telecommunications. 2022.

  • China National Scholarship. Ministry of Education of P.R. China. 2020.

  • First Prize in National College Student Mathematics Competition. Chinese Mathematics League. 2018.



Service

  • Reviewer: EMNLP2022, ACL2023, EMNLP2023, ARR




Design and source code from Jon Barron's website