Variational Open-Domain Question Answering

Research output: Contribution to journalConference articleResearchpeer-review

Standard

Variational Open-Domain Question Answering. / Liévin, Valentin; Motzfeldt, Andreas Geert; Jensen, Ida Riis; Winther, Ole.

In: Proceedings of Machine Learning Research, Vol. 202, 2023, p. 20950-20977.

Research output: Contribution to journalConference articleResearchpeer-review

Harvard

Liévin, V, Motzfeldt, AG, Jensen, IR & Winther, O 2023, 'Variational Open-Domain Question Answering', Proceedings of Machine Learning Research, vol. 202, pp. 20950-20977. <https://proceedings.mlr.press/v202/>

APA

Liévin, V., Motzfeldt, A. G., Jensen, I. R., & Winther, O. (2023). Variational Open-Domain Question Answering. Proceedings of Machine Learning Research, 202, 20950-20977. https://proceedings.mlr.press/v202/

Vancouver

Liévin V, Motzfeldt AG, Jensen IR, Winther O. Variational Open-Domain Question Answering. Proceedings of Machine Learning Research. 2023;202:20950-20977.

Author

Liévin, Valentin ; Motzfeldt, Andreas Geert ; Jensen, Ida Riis ; Winther, Ole. / Variational Open-Domain Question Answering. In: Proceedings of Machine Learning Research. 2023 ; Vol. 202. pp. 20950-20977.

Bibtex

@inproceedings{69332facf10f462fb052bcc15d938f99,
title = "Variational Open-Domain Question Answering",
abstract = "Retrieval-augmented models have proven to be effective in natural language processing tasks, yet there remains a lack of research on their optimization using variational inference. We introduce the Variational Open-Domain (VOD) framework for end-to-end training and evaluation of retrieval-augmented models, focusing on open-domain question answering and language modelling. The VOD objective, a self-normalized estimate of the R{\'e}nyi variational bound, approximates the task marginal likelihood and is evaluated under samples drawn from an auxiliary sampling distribution (cached retriever and/or approximate posterior). It remains tractable, even for retriever distributions defined on large corpora. We demonstrate VOD's versatility by training reader-retriever BERT-sized models on multiple-choice medical exam questions. On the MedMCQA dataset, we outperform the domain-tuned Med-PaLM by +5.3% despite using 2.500× fewer parameters. Our retrieval-augmented BioLinkBERT model scored 62.9% on the MedMCQA and 55.0% on the MedQA-USMLE. Last, we show the effectiveness of our learned retriever component in the context of medical semantic search.",
author = "Valentin Li{\'e}vin and Motzfeldt, {Andreas Geert} and Jensen, {Ida Riis} and Ole Winther",
note = "Publisher Copyright: {\textcopyright} 2023 Proceedings of Machine Learning Research. All rights reserved.; 40th International Conference on Machine Learning, ICML 2023 ; Conference date: 23-07-2023 Through 29-07-2023",
year = "2023",
language = "English",
volume = "202",
pages = "20950--20977",
journal = "Proceedings of Machine Learning Research",
issn = "2640-3498",

}

RIS

TY - GEN

T1 - Variational Open-Domain Question Answering

AU - Liévin, Valentin

AU - Motzfeldt, Andreas Geert

AU - Jensen, Ida Riis

AU - Winther, Ole

N1 - Publisher Copyright: © 2023 Proceedings of Machine Learning Research. All rights reserved.

PY - 2023

Y1 - 2023

N2 - Retrieval-augmented models have proven to be effective in natural language processing tasks, yet there remains a lack of research on their optimization using variational inference. We introduce the Variational Open-Domain (VOD) framework for end-to-end training and evaluation of retrieval-augmented models, focusing on open-domain question answering and language modelling. The VOD objective, a self-normalized estimate of the Rényi variational bound, approximates the task marginal likelihood and is evaluated under samples drawn from an auxiliary sampling distribution (cached retriever and/or approximate posterior). It remains tractable, even for retriever distributions defined on large corpora. We demonstrate VOD's versatility by training reader-retriever BERT-sized models on multiple-choice medical exam questions. On the MedMCQA dataset, we outperform the domain-tuned Med-PaLM by +5.3% despite using 2.500× fewer parameters. Our retrieval-augmented BioLinkBERT model scored 62.9% on the MedMCQA and 55.0% on the MedQA-USMLE. Last, we show the effectiveness of our learned retriever component in the context of medical semantic search.

AB - Retrieval-augmented models have proven to be effective in natural language processing tasks, yet there remains a lack of research on their optimization using variational inference. We introduce the Variational Open-Domain (VOD) framework for end-to-end training and evaluation of retrieval-augmented models, focusing on open-domain question answering and language modelling. The VOD objective, a self-normalized estimate of the Rényi variational bound, approximates the task marginal likelihood and is evaluated under samples drawn from an auxiliary sampling distribution (cached retriever and/or approximate posterior). It remains tractable, even for retriever distributions defined on large corpora. We demonstrate VOD's versatility by training reader-retriever BERT-sized models on multiple-choice medical exam questions. On the MedMCQA dataset, we outperform the domain-tuned Med-PaLM by +5.3% despite using 2.500× fewer parameters. Our retrieval-augmented BioLinkBERT model scored 62.9% on the MedMCQA and 55.0% on the MedQA-USMLE. Last, we show the effectiveness of our learned retriever component in the context of medical semantic search.

M3 - Conference article

AN - SCOPUS:85172434053

VL - 202

SP - 20950

EP - 20977

JO - Proceedings of Machine Learning Research

JF - Proceedings of Machine Learning Research

SN - 2640-3498

T2 - 40th International Conference on Machine Learning, ICML 2023

Y2 - 23 July 2023 through 29 July 2023

ER -

ID: 372185038