Can Reasoning Models Enhance Embedding Models?
Project Description
Recent embedding models are initialized from large language model (LLMs), enjoying the massive knowledge encoded in LLMs weights. However, prior works only consider non-reasoning model, and they didn’t study the impact of employing a reasoning model (trained with Reinforcement Learning with Verifiable Reward (RLVR)) as a base model. In this work, we want to answer one question: can a reasoning model lead to a better embedding model? To answer this question, we will explore whether reasoning capabilities can improve the semantic richness and performance of embedding models, by investigating RLVR’s impact on embedding space, and training multiple embedding models to compare the performance difference.
Supervisor
SONG Yangqiu
Quota
2
Course type
UROP1100
UROP2100
UROP3100
UROP3200
UROP4100
Applicant's Roles
Working together with a MPhil student on task formulation, designing experiments, analyzing results, and writing research papers.
Applicant's Learning Objectives
1. Have hands-on experience in playing with reasoning model, embedding model, and model training & evaluation.
2. Learn how to research with them for diverse reasoning scenarios.
3. Gain deep understanding of recent advancements in reasoning model and embedding model.
2. Learn how to research with them for diverse reasoning scenarios.
3. Gain deep understanding of recent advancements in reasoning model and embedding model.
Complexity of the project
Challenging