Show simple item record

dc.contributor.advisorAgirre Bengoa, Eneko ORCID
dc.contributor.advisorAzkune Galparsoro, Gorka
dc.contributor.authorDomínguez Becerril, Carlos
dc.date.accessioned2023-06-30T15:15:25Z
dc.date.available2023-06-30T15:15:25Z
dc.date.issued2023-06-30
dc.identifier.urihttp://hdl.handle.net/10810/61836
dc.description.abstractNowadays, to tackle the open domain Question Answering (QA) problem, a neural architecture with two main components is usually used: the information retriever, whose task is to search for the most relevant documents with respect to the question, and the reader, which given the question and the extracted documents that serve as context, generates the appropriate answer. In this project, we propose to investigate open domain QA, with a special focus on the first component of the architecture, that is, the information retrieval component. We want to train several dense neural retrievers in an unsupervised manner by generating questions from the documents using Large Language Models (LLM). Currently, most LLMs provide several checkpoints with different amounts of parameters, and we want to use those checkpoints to generate questions, train a dense neural retriever system for each LLM checkpoint, and finally, compare if the generated questions have any influence in the performance of the systems. The developed system must be able to search for the necessary information in external sources, usually organized in text documents. For that purpose, the BEIR benchmark \cite{BEIR} will be used in a zero-shot manner to test the performance. As the result of this exploration, we found that using LLMs to generate questions can be helpful in order to train information retrieval systems as it achieves similar performance to supervised systems. More concretely, we found that: (i) the more parameters the LLM has the more performance we obtain, (ii) using sampling to generate questions can further increase the performance, and (iii) generating more questions using a smaller language model is not worth it as a checkpoint with more parameters can do a better job. Moreover, we found that using different prompts and/or domain adaptation on a specific dataset can improve the performance slightly.es_ES
dc.language.isoenges_ES
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectartificial Intelligencees_ES
dc.subjectdeep learninges_ES
dc.subjectnatural language processinges_ES
dc.subjectdense passage retrievales_ES
dc.subjectquestion generationes_ES
dc.subjectunsupervised traininges_ES
dc.titleUnsupervised information retrieval using large language modelses_ES
dc.typeinfo:eu-repo/semantics/masterThesis
dc.date.updated2023-02-09T11:17:24Z
dc.language.rfc3066es
dc.rights.holder© 2023, el autor
dc.contributor.degreeMáster Universitario en Análisis y Procesamiento del Lenguaje
dc.contributor.degreeHizkuntzaren Azterketa eta Prozesamendua Unibertsitate Masterra
dc.identifier.gaurregister128881-879042-05es_ES
dc.identifier.gaurassign147928-879042es_ES


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record