Global and Local Context Fusion in Heterogeneous Graph Neural Network for Summarizing Lengthy Scientific Documents

Muhammad Umair, Atif Khan, Fasee Ullah, Atef Masmoudi, Muhammad Faheem*

*Corresponding author for this work

Research output: Contribution to journalArticleScientificpeer-review

Abstract

The primary objective of text summarization is to condense a document’s length while preserving its essential content. Extractive summarization methods are commonly used due to their effectiveness and straightforward presentation. However, a significant challenge lies in segmenting documents into distinct concepts and understanding how sentences interact, especially in complex materials such as scientific articles. This process entails identifying relationships between sentences and determining the most significant and informative content within extensive text collections. Traditional techniques often utilize pre-trained models like BERT, known for their ability to capture word context. Nonetheless, these models have limitations, including constrained input lengths and the computational intensity of self-attention mechanisms, which hinder their effectiveness in processing large-scale scientific texts. To address these challenges, we propose a computationally efficient Heterogeneous Graph Neural Network (HGNN) for the extractive summarization of lengthy scientific texts. This framework combines GloVe embeddings with Convolutional Neural Networks (CNNs) and Bidirectional Long Short-Term Memory (BiLSTM) encoders. GloVe offers simple yet effective word embeddings, CNNs focus on capturing local word structures, and BiLSTMs identify long-range dependencies, allowing for flexible encoding of extensive texts. For global context and topic modeling, we utilize an enhanced version of Latent Dirichlet Allocation (LDA) to retain essential document attributes. In this model, words, sentences, and topics are represented as nodes in a heterogeneous graph, with TFIDF values illustrating the relationships between edges. The graph is processed using a Graph Attention Network (GAT), which refines node representations by integrating both local and global information. This study represents the first instance of combining LDA with CNN and BiLSTM encoders in a Graph Attention-based model for summarizing scientific texts. Experimental results demonstrate that the proposed framework outperforms both BERT-based and non-BERT approaches on publicly available datasets from arXiv and PubMed.

Original languageEnglish
Pages (from-to)53433-53447
JournalIEEE Access
Volume13
DOIs
Publication statusPublished - 2025
MoE publication typeA1 Journal article-refereed

Keywords

  • BiLSTM
  • CNN
  • DOCUMENT SUMMARIZATION
  • GATs
  • GloVe
  • LDA
  • SCIENTIFIC PAPERS LONG
  • TF-IDF

Fingerprint

Dive into the research topics of 'Global and Local Context Fusion in Heterogeneous Graph Neural Network for Summarizing Lengthy Scientific Documents'. Together they form a unique fingerprint.

Cite this