In the past few years, the field of text simplification has been dominated by supervised learning approaches thanks to the appearance of large parallel datasets such as Wikilarge and Newsela. However, these datasets suffer from sentence pairs with factuality errors which compromise the models’ performance. So, we proposed a model-independent factuality error detection mechanism, considering bad simplification and bad alignment, to refine the Wikilarge dataset through reducing the weight of these samples during training. We demonstrated that this approach improved the performance of the state-of-the-art text simplification model TST5 by an FKGL reduction of 0.33 and 0.29 on the TurkCorpus and ASSET testing datasets respectively. Our study illustrates the impact of erroneous samples in TS datasets and highlights the need for automatic methods to improve their quality.
Ashwin DevarajWilliam P. SheffieldByron WallaceJunyi Jessy Li
Devaraj, Ashwin0000-0001-5571-0681
Max SchwarzerTeerapaun TanprasertDavid Kauchak
Max SchwarzerTeerapaun TanprasertDavid Kauchak
Hudgins, KyleOnder, SerraTang, RongZiyu Zhou