This page is on the 2022 long paper
on Conversational Question Answering on Heterogeneous Sources. We provide an extended video
as an introduction to the work (a shorter video is also available below).
GitHub link to CONVINSE code
Directly download CONVINSE code
Conversational question answering (ConvQA) tackles sequential information needs where contexts in follow-up questions are left implicit. Current ConvQA systems operate over homogeneous sources of information: either a knowledge base (KB), or a text corpus, or a collection of tables. This paper addresses the novel issue of jointly tapping into all of these together, this way boosting answer coverage and confidence. We present CONVINSE, an end-to-end pipeline for ConvQA over heterogeneous sources, operating in three stages: i) learning an explicit structured representation of an incoming question and its conversational context, ii) harnessing this frame-like representation to uniformly capture relevant evidences from KB, text, and tables, and iii) running a fusion-in-decoder model to generate the answer. We construct and release the first benchmark, ConvMix, for ConvQA over heterogeneous sources, comprising 3000 real-user conversations with 16000 questions, along with entity annotations, completed question utterances, and question paraphrases. Experiments demonstrate the viability and advantages of our method, compared to state-of-the-art baselines.
For feedback and clarifications, please contact: Philipp Christmann
(pchristm AT mpi HYPHEN inf DOT mpg DOT de), Rishiraj Saha Roy
(rishiraj AT mpi HYPHEN inf DOT mpg DOT de) or Gerhard Weikum
(weikum AT mpi HYPHEN inf DOT mpg DOT de).
To know more about our group, please visit https://www.mpi-inf.mpg.de/departments/databases-and-information-systems/research/question-answering/
Training Set (1680 Conversations)
Dev Set (560 Conversations)
The ConvMix benchmark is licensed under a Test Set (760 Conversations)Creative Commons Attribution 4.0 International License
Christmann et al. '22
Question Resolution + BM25 + FiD
Question Rewriting + BM25 + FiD
What do conversations in ConvMix look like?
The sources in square brackets are the ones the respective answer can be found in.
How was ConvMix created?
benchmark was created by real humans, and we tried to ensure that the collected data is as natural as possible. Overall, it contains 3,000 conversations with 16,000 unique questions. Master crowdworkers
on Amazon Mechanical Turk (AMT) selected an entity of interest in a specific domain, and then started issuing conversational questions on this entity, potentially drifting to other topics of interest throughout the course of the conversation. By letting users choose the entities themselves, we aimed to ensure that they are more interested into the topics the conversations are based on.
After writing a question, users were asked to find the answer in eithers Wikidata
, Wikipedia text
, a Wikipedia table
or a Wikipedia infobox
, whatever they find more natural for the specific question at hand.
Since Wikidata requires some basic understanding of knowledge bases, we provided video guidelines that illustrated how Wikidata can be used for detecting answers, following an example conversation.
We provide not only the question and answer, but also the answer source the user found the answer in, a paraphrase, a completed question, and question entities.
For further details on ConvMix, please refer to the paper
"Conversational Question Answering on Heterogeneous Sources"
, Philipp Christmann, Rishiraj Saha Roy, and Gerhard Weikum. In SIGIR '22
, Madrid, Spain, 11 - 15 July 2022.
] [Extended video
] [User study