Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about bge-en-icl paper #1213

Open
whybe-choi opened this issue Nov 12, 2024 · 1 comment
Open

Question about bge-en-icl paper #1213

whybe-choi opened this issue Nov 12, 2024 · 1 comment

Comments

@whybe-choi
Copy link

Dear Authors,

Firstly, thank you for your great work, "Making Text Embedder Few-Shot Learners". It was very interesting to see how you improved the performance of text embedding by leveraging the intrinsic capabilities of LLMs!

While studying for the paper, I had a question. In Section 3.1, you note that embedding models have a limited ability to follow unseen embedding task instructions and conduct complex retrieval tasks. then you explore whether embedding models can be enhanced by leveraging ICL.

When doing the few-shot contrastive training, it was my understanding that the same instruction was applied to “task_definition” for each specific dataset. In this respect, I think it is not much different from previous studies that the model was trained with limited instructions for the dataset used in the training process, so I would like to know more about the difference between previous studies and this paper that utilizes ICL.

Thank you for your time and for contributing such valuable research to the community.

@545999961
Copy link
Collaborator

Thank you for your attention to our work. ICL does not provide additional semantic information in the task definition part. Instead, it introduces in-context examples to help better understand the task intent. By using in-context examples, ICL can achieve better results on out-of-domain tasks by simply providing task-relevant examples, compared to using only the task definition.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants