-
Notifications
You must be signed in to change notification settings - Fork 130
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Release of training code for QKConv #179
Comments
Thanks for your attention to this work! |
@christineaa I have a one more question regarding the experiment in the QKConv paper. And similarly for the test dataset - did you use the conversation level or question-answer level? Thank you! |
We used the question-answer pairs as the training/dev/test dataset, with 60.4K, 3.1K, and 16.4K samples respectively. |
Hi, @christineaa . Thanks for your nice work. I have one more question: how should I build the BM25 index for QRecc task. I notice you post a link to the ml-qrecc repo. Whether should I download the webpages from both the Common Crawl and the Wayback Machine and build the BM25 index? |
Thanks for your attention to this work! |
Thanks for your reply. |
@christineaa I have further questions regarding training and evaluation of QKConv model on QReCC dataset.
|
Also, when you report Table2, did you exlcude test examples which do not have gold knowledge ? |
@robinsongh381 Thanks for your attention to this work!
|
@christineaa Thank you for kind response. I have a further question for Question 3. The absence of gold knowledge indicates that the essential and required piece of information does not exist within the knowledge pool and hence factually correct and knowledg-grounded response cannot be obtained. For this reason, I have found that previous works on QReCC evaluation, such as DPR-IHN[1], and CONQRR[2], have excluded such cases (i.e., examples without gold-knowledge annotation) in their evaluation. What is your opinion on this ? Thank you [1] Saving Dense Retriever from Shortcut Dependency in Conversational Search [2] CONQRR: Conversational Query Rewriting for Retrieval with Reinforcement Learning |
@robinsongh381 However, DPR-IHN and CONQRR excluding samples without golden knowledge are another case. They present knowledge selection Recall metrics as their main results, and Recall metrics cannot be applied without golden knowledge. |
@christineaa Thanks for sharing nice work !
Do you have any plans to release the training code ?
The text was updated successfully, but these errors were encountered: