Open Access
ARTICLE
Prompt-Guided Dialogue State Tracking with GPT-2 and Graph Attention
1 School of Computer Science and Engineering, Southeast University, Nanjing, 211189, China
2 Department of AI and Data Science, Sejong University, Seoul, 05006, Republic of Korea
3 Department of Electronics and Communication Engineering, National Institute of Technology, Rourkela, 769008, India
4 Department of Computer Science, Shaheed Benazir Bhutto University, Sheringal, 18050, Pakistan
5 Department of Computer Science, University of Engineering and Technology, Mardan, 23200, Pakistan
6 School of Computing, Gachon University, Seongnam, 13120, Republic of Korea
7 Department of Informatics, Aristotle University of Thessaloniki, Thessaloniki, 54124, Greece
* Corresponding Authors: Jawad Khan. Email: ; Yeong Hyeon Gu. Email:
Computers, Materials & Continua 2025, 85(3), 5451-5468. https://doi.org/10.32604/cmc.2025.069134
Received 15 June 2025; Accepted 29 August 2025; Issue published 23 October 2025
Abstract
Dialogue State Tracking (DST) is a critical component of task-oriented spoken dialogue systems (SDS), tasked with maintaining an accurate representation of the conversational state by predicting slots and their corresponding values. Recent advances leverage Large Language Models (LLMs) with prompt-based tuning to improve tracking accuracy and efficiency. However, these approaches often incur substantial computational and memory overheads and typically address slot extraction implicitly within prompts, without explicitly modeling the complex dependencies between slots and values. In this work, we propose PUGG, a novel DST framework that constructs schema-driven prompts to fine-tune GPT-2 and utilizes its tokenizer to implement a memory encoder. PUGG explicitly extracts slot values via GPT-2 and employs Graph Attention Networks (GATs) to model and reason over the intricate relationships between slots and their associated values. We evaluate PUGG on four publicly available datasets, where it achieves state-of-the-art performance across multiple evaluation metrics, highlighting its robustness and generalizability in diverse conversational scenarios. Our results indicate that the integration of GPT-2 substantially reduces model complexity and memory consumption by streamlining key processes. Moreover, prompt tuning enhances the model’s flexibility and precision in extracting relevant slot-value pairs, while the incorporation of GATs facilitates effective relational reasoning, leading to improved dialogue state representations.Keywords
Cite This Article
Copyright © 2025 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF
Downloads
Citation Tools