JOURNAL ARTICLE

Abstractive Summarization of Korean Legal Cases using Pre-trained Language Models

Jiyoung YoonMuhammad JunaidSajid AliJongwuk Lee

Year: 2022 Journal:   2022 16th International Conference on Ubiquitous Information Management and Communication (IMCOM) Pages: 1-7

Abstract

AI technology in the legal domain has developed at a rapid pace around the world, but not much research is being conducted in the Korean legal field due to barriers of language and the high level of expertise required. We first attempt abstractive summarization of Korean legal decision text and publicly release our collected dataset. We utilize two pretrained language models, i.e., BERT2BERT and BART, for our task. They are based on the encoder-decoder approach under transformer architecture. While BERT2BERT is pre-trained with BERT on both the encoder and decoder, BART combines BERT and GPT as the encoder and the decoder. We then evaluate the baseline models and show that, despite the difference in language style, the high-quality summary was generated using applied models. We also show that pre-training using both autoencoder and autoregressive method makes better performance than using solely denoising autoencoder.

Keywords:
Automatic summarization Computer science Autoencoder Encoder Artificial intelligence Natural language processing Autoregressive model Transformer Pace Language model Baseline (sea) Speech recognition Machine learning Deep learning Engineering

Metrics

13
Cited By
0.00
FWCI (Field Weighted Citation Impact)
28
Refs
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

© 2026 ScienceGate Book Chapters — All rights reserved.