TY - JOUR
T1 - Challenges in Domain-Specific Abstractive Summarization and How to Overcome Them
AU - Afzal, Anum
AU - Vladika, Juraj
AU - Braun, Daniel
AU - Matthes, Florian
N1 - Publisher Copyright:
© 2023 by SCITEPRESS – Science and Technology Publications, Lda.
PY - 2023
Y1 - 2023
N2 - Large Language Models work quite well with general-purpose data and many tasks in Natural Language Processing. However, they show several limitations when used for a task such as domain-specific abstractive text summarization. This paper identifies three of those limitations as research problems in the context of abstractive text summarization: 1) Quadratic complexity of transformer-based models with respect to the input text length; 2) Model Hallucination, which is a model’s ability to generate factually incorrect text; and 3) Domain Shift, which happens when the distribution of the model’s training and test corpus is not the same. Along with a discussion of the open research questions, this paper also provides an assessment of existing state-of-the-art techniques relevant to domain-specific text summarization to address the research gaps.
AB - Large Language Models work quite well with general-purpose data and many tasks in Natural Language Processing. However, they show several limitations when used for a task such as domain-specific abstractive text summarization. This paper identifies three of those limitations as research problems in the context of abstractive text summarization: 1) Quadratic complexity of transformer-based models with respect to the input text length; 2) Model Hallucination, which is a model’s ability to generate factually incorrect text; and 3) Domain Shift, which happens when the distribution of the model’s training and test corpus is not the same. Along with a discussion of the open research questions, this paper also provides an assessment of existing state-of-the-art techniques relevant to domain-specific text summarization to address the research gaps.
KW - Domain-Adaptation of Language Models
KW - Efficient Transformers
KW - Model Hallucination
KW - Natural Language Generation Evaluation
KW - Natural Language Processing
KW - Text Summarization
UR - http://www.scopus.com/inward/record.url?scp=85177733870&partnerID=8YFLogxK
U2 - 10.5220/0011744500003393
DO - 10.5220/0011744500003393
M3 - Conference article
AN - SCOPUS:85177733870
SN - 2184-3589
VL - 3
SP - 682
EP - 689
JO - International Conference on Agents and Artificial Intelligence
JF - International Conference on Agents and Artificial Intelligence
T2 - 15th International Conference on Agents and Artificial Intelligence, ICAART 2023
Y2 - 22 February 2023 through 24 February 2023
ER -