
@Article{cmc.2025.069134,
AUTHOR = {Muhammad Asif Khan, Dildar Hussain, Bhuyan Kaibalya Prasad, Irfan Ullah, Inayat Khan, Jawad Khan, Yeong Hyeon Gu, Pavlos Kefalas},
TITLE = {Prompt-Guided Dialogue State Tracking with GPT-2 and Graph Attention},
JOURNAL = {Computers, Materials \& Continua},
VOLUME = {85},
YEAR = {2025},
NUMBER = {3},
PAGES = {5451--5468},
URL = {http://www.techscience.com/cmc/v85n3/64190},
ISSN = {1546-2226},
ABSTRACT = {Dialogue State Tracking (DST) is a critical component of task-oriented spoken dialogue systems (SDS), tasked with maintaining an accurate representation of the conversational state by predicting slots and their corresponding values. Recent advances leverage Large Language Models (LLMs) with prompt-based tuning to improve tracking accuracy and efficiency. However, these approaches often incur substantial computational and memory overheads and typically address slot extraction implicitly within prompts, without explicitly modeling the complex dependencies between slots and values. In this work, we propose <b>PUGG</b>, a novel DST framework that constructs schema-driven prompts to fine-tune GPT-2 and utilizes its tokenizer to implement a memory encoder. PUGG explicitly extracts slot values via GPT-2 and employs Graph Attention Networks (GATs) to model and reason over the intricate relationships between slots and their associated values. We evaluate PUGG on four publicly available datasets, where it achieves state-of-the-art performance across multiple evaluation metrics, highlighting its robustness and generalizability in diverse conversational scenarios. Our results indicate that the integration of GPT-2 substantially reduces model complexity and memory consumption by streamlining key processes. Moreover, prompt tuning enhances the model’s flexibility and precision in extracting relevant slot-value pairs, while the incorporation of GATs facilitates effective relational reasoning, leading to improved dialogue state representations.},
DOI = {10.32604/cmc.2025.069134}
}



