Using Pre-trained Language Models for Abstractive DBpedia Summarization: A Comparative Study

H.M.A. Zahera, F. Vitiugin, M. Sherif, C. Castillo, A.-C. Ngonga Ngomo, in: SEMANTiCS, 2023.

Download
No fulltext has been uploaded.
Conference Paper | English
Abstract
Purpose: This study addresses the limitations of current short abstracts of DBpedia entities, which often lack a comprehensive overview due to their creating method (i.e., selecting the first two-three sentences from the full DBpedia abstracts). Methodology: We leverage pre-trained language models to generate abstractive summaries of DBpedia abstracts in six languages (English, French, German, Italian, Spanish, and Dutch). We performed several experiments to assess the quality of generated summaries by language models. In particular, we evaluated the generated summaries using human judgments and automated metrics (Self-ROUGE and BERTScore). Additionally, we studied the correlation between human judgments and automated metrics in evaluating the generated summaries under different aspects: informativeness, coherence, conciseness, and fluency. Findings: Pre-trained language models generate summaries more concise and informative than existing short abstracts. Specifically, BART-based models effectively overcome the limitations of DBpedia short abstracts, especially for longer ones. Moreover, we show that BERTScore and ROUGE-1 are reliable metrics for assessing the informativeness and coherence of the generated summaries with respect to the full DBpedia abstracts. We also find a negative correlation between conciseness and human ratings. Furthermore, fluency evaluation remains challenging without human judgment. Value: This study has significant implications for various applications in machine learning and natural language processing that rely on DBpedia resources. By providing succinct and comprehensive summaries, our approach enhances the quality of DBpedia abstracts and contributes to the semantic web community
Publishing Year
Proceedings Title
SEMANTiCS
Conference
SEMANTiCS 2023
Conference Location
Leipzig, Germany
Conference Date
2023-09-20 – 2023-09-23
LibreCat-ID

Cite this

Zahera HMA, Vitiugin F, Sherif M, Castillo C, Ngonga Ngomo A-C. Using Pre-trained Language Models for Abstractive DBpedia Summarization: A Comparative Study. In: SEMANTiCS. ; 2023.
Zahera, H. M. A., Vitiugin, F., Sherif, M., Castillo, C., & Ngonga Ngomo, A.-C. (2023). Using Pre-trained Language Models for Abstractive DBpedia Summarization: A Comparative Study. SEMANTiCS. SEMANTiCS 2023, Leipzig, Germany.
@inproceedings{Zahera_Vitiugin_Sherif_Castillo_Ngonga Ngomo_2023, title={Using Pre-trained Language Models for Abstractive DBpedia Summarization: A Comparative Study}, booktitle={SEMANTiCS}, author={Zahera, Hamada Mohamed Abdelsamee and Vitiugin, Fedor and Sherif, Mohamed and Castillo, Carlos and Ngonga Ngomo, Axel-Cyrille}, year={2023} }
Zahera, Hamada Mohamed Abdelsamee, Fedor Vitiugin, Mohamed Sherif, Carlos Castillo, and Axel-Cyrille Ngonga Ngomo. “Using Pre-Trained Language Models for Abstractive DBpedia Summarization: A Comparative Study.” In SEMANTiCS, 2023.
H. M. A. Zahera, F. Vitiugin, M. Sherif, C. Castillo, and A.-C. Ngonga Ngomo, “Using Pre-trained Language Models for Abstractive DBpedia Summarization: A Comparative Study,” presented at the SEMANTiCS 2023, Leipzig, Germany, 2023.
Zahera, Hamada Mohamed Abdelsamee, et al. “Using Pre-Trained Language Models for Abstractive DBpedia Summarization: A Comparative Study.” SEMANTiCS, 2023.

Export

Marked Publications

Open Data LibreCat

Search this title in

Google Scholar