On Monotonic Aggregation for Open-domain QA

Sang Eun Han, Yeonseok Jeong, Seung Won Hwang, Kyungjae Lee

Research output: Contribution to journalConference articlepeer-review

Abstract

Question answering (QA) is a critical task for speech-based retrieval from knowledge sources, by sifting only the answers without requiring to read supporting documents. Specifically, open-domain QA aims to answer user questions on unrestricted knowledge sources. Ideally, adding a source should not decrease the accuracy, but we find this property (denoted as “monotonicity”) does not hold for current state-of-the-art methods. We identify the cause, and based on that we propose Judge-Specialist framework. Our framework consists of (1) specialist retrievers/readers to cover individual sources, and (2) judge, a dedicated language model to select the final answer. Our experiments show that our framework not only ensures monotonicity, but also outperforms state-of-the-art multi-source QA methods on Natural Questions. Additionally, we show that our models robustly preserve the monotonicity against noise from speech recognition. We publicly release our code and setting.

Original languageEnglish
Pages (from-to)3432-3436
Number of pages5
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume2023-August
DOIs
StatePublished - 2023
Event24th International Speech Communication Association, Interspeech 2023 - Dublin, Ireland
Duration: 20 Aug 202324 Aug 2023

Keywords

  • open domain QA
  • QA from speech

Fingerprint

Dive into the research topics of 'On Monotonic Aggregation for Open-domain QA'. Together they form a unique fingerprint.

Cite this