Mining Implicit Relevance Feedback from User Behavior for Web Question Answering
Linjun Shou: STCA NLP Group Microsoft Beijing ; Shining Bo: School of Information and Communication Engineering BUPT ; Feixiang Cheng: STCA NLP Group Microsoft Beijing ; Ming Gong: STCA NLP Group Microsoft Beijing ; Jian Pei: School of Computing Science Simon Fraser University ; Daxin Jiang: STCA NLP Group Microsoft Beijing
Training and refreshing a web-scale Question Answering (QA) system for a multi-lingual commercial search engine often requires a huge amount of training examples. One principled idea is to mine implicit relevance feedback from user behavior recorded in search engine logs. All previous works on mining implicit relevance feedback target at relevance of web documents rather than passages. Due to several unique characteristics of QA tasks, the existing user behavior models for web documents cannot be applied to infer passage relevance. In this paper, we make the first study to explore the correlation between user behavior and passage relevance, and propose a novel approach for mining training data for Web QA. We conduct extensive experiments on four test datasets and the results show our approach significantly improves the accuracy of passage ranking without extra human labeled data. In practice, this work has proved effective to substantially reduce the human labeling cost for the QA service in a global commercial search engine, especially for languages with low resources. Our techniques have been deployed in multi-language services.
How can we assist you?
We'll be updating the website as information becomes available. If you have a question that requires immediate attention, please feel free to contact us. Thank you!
Please enter the word you see in the image below: