Dialogue State Tracking (DST) is an essential part of task-oriented dialogue systems. Many existing methods try to utilize external dialogue datasets to improve the performance of DST models. Instead of previous methods, in this paper, we propose Label-Aware Auxiliary Learning for DST (LAL-DST) which focuses on exploiting the abundant internal information of the target DST dataset to improve the performance. We design label-aware auxiliary tasks, in which we apply noising functions to either the dialogue history or the belief state label and take the concatenation of them as input. The goal of each task is to restore the corrupted context. During the training process, we first further train the large pre-trained language model on the auxiliary tasks, then fine-tune it on DST. Through the experimental results, we empirically show the effect of LAL-DST by the performance improvements it brings to MultiWOZ2.0 and WOZ.
Fanghua YeYong FengEmine Yılmaz
Sijie FengHaoxiang SuHongyan XieDi WuHao HuangWushour Silamu
Zhi ChenLu ChenYanbin ZhaoSu ZhuKai Yu
Pradeep Kumar MS. BoseDanya SG GokulrajPrabhu Ragavendiran S DMaheswaran N