ADA Library Digital Repository

Exploring the Use of Knowledge Graphs in Enhancing Natural Language Understanding for Conversational AI

Show simple item record

dc.contributor.author Guliyeva, Fidan
dc.date.accessioned 2025-10-27T05:01:48Z
dc.date.available 2025-10-27T05:01:48Z
dc.date.issued 2025
dc.identifier.uri http://hdl.handle.net/20.500.12181/1498
dc.description.abstract Recent advancements in large language models have significantly contributed to the capabilities of conversational AI systems by empowering coherent, context-aware dialogues across a wide range of topics. Although large language models showcase impressive fluency and verbosity, many studies show that these models are prone to hallucinations, which is highly problematic, especially in domains where factual reliability is required. This research investigates the use of a structured knowledge base in the efforts to minimize hallucination in multi-hop question-answering tasks with a focus on a low-resource language such as Azerbaijani. As a result of the investigations, we choose and adopt the Graph-Constrained Reasoning framework, which integrates Knowledge Graph structures directly into the decoding process of LLMs. This framework enforces graph-based constraints while generating reasoning paths by constructing a KG-Trie from the given dataset. As a preliminary step, we translated the WebQSP and ComplexWebQuestions datasets into the Azerbaijani language. As a part of the chosen framework, we fine-tuned multilingual mT5 models (small, base, and large variants) on the Azerbaijani-translated datasets and evaluated their performance against a monolingual English baseline model, Qwen2-0.5B. Initial results indicated underfitting due to limited dataset size. This issue motivated the development of a data augmentation strategy that generates multiple reasoning paths from all combinations of question and answer entities. The augmentation significantly expanded the training corpus and led to considerable performance improvements across all mT5 models. The final evaluation shows that the mT5-large variant achieved an F1 score of 75.2 and a HIT score of 67.0 on the WebQSP dataset, substantially closing the performance gap with the Qwen2-0.5B baseline. Despite these gains, the mT5 models could not surpass the Qwen2-0.5B model in F1 score. This study shows that structured grounding, achieved through graph-constrained decoding, can improve factual reliability in multilingual large language models. en_US
dc.language.iso en en_US
dc.publisher ADA University en_US
dc.rights Attribution-NonCommercial-NoDerivs 3.0 United States *
dc.rights.uri http://creativecommons.org/licenses/by-nc-nd/3.0/us/ *
dc.subject Large language models (Computer science) en_US
dc.subject Natural language processing (Computer science) en_US
dc.subject Question answering (Information retrieval) en_US
dc.subject Machine learning -- Multilingualism en_US
dc.subject Computational linguistics -- Azerbaijani language en_US
dc.title Exploring the Use of Knowledge Graphs in Enhancing Natural Language Understanding for Conversational AI en_US
dc.type Thesis en_US
dcterms.accessRights Absolute Embargo (No access without the author's permission)


Files in this item

Files Size Format View

There are no files associated with this item.

The following license files are associated with this item:

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-NoDerivs 3.0 United States Except where otherwise noted, this item's license is described as Attribution-NonCommercial-NoDerivs 3.0 United States

Search ADA LDR


Advanced Search

Browse

My Account