By leveraging a multilingual language model, we show that CTRLSum [1], an abstractive summarization approach that can be controlled by keywords, improves baseline summarization system in four languages: English, Indonesian, Spanish, and French by 1.57 in terms of average ROUGE-1, with the Indonesian model achieving state-of-the-art results. We further provide novel analysis about the importance of keywords fed to CTRLSum which (1) shows hypothetical upper-bound results that outperform the state-of-the-art in all four languages by a large margin and (2) provides natural direction for future work to improve CTRLSum by improving the keyword prediction model.
Aiom Minnette MitriGoutam SahaSaralin A. LyngdohArnab Kumar Maji
Angela FanDavid GrangierMichael Auli
Mwnthai NarzaryPranav Kumar Singh
Simon MilleBallesteros MiguelBurga AliciaCasamayor GerardWilliam Leo