E2EG: End-to-End Node Classification Using Graph Topology and Text-based Node Attributes

Open Access
Authors
  • T.A. Dinh
  • J. den Boef
  • J. Cornelisse
  • P. Groth ORCID logo
Publication date 2023
Host editors
  • J. Wang
  • Y. He
  • T.N. Dinh
  • C. Grant
  • M. Qiu
  • W. Pedrycz
Book title 23rd IEEE International Conference on Data Mining Workshops
Book subtitle 1-4 December 2023, Shanghai, China : proceedings
ISBN
  • 9798350381658
ISBN (electronic)
  • 9798350381641
Series ICDMW
Event 23rd IEEE International Conference on Data Mining Workshops
Pages (from-to) 1084-1091
Number of pages 8
Publisher Los Alamitos, CA: IEEE Computer Society
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Node classification utilizing text-based node attributes has many real-world applications, ranging from prediction of paper topics in academic citation graphs to classification of user characteristics in social media networks. State-of-the-art node classification frameworks, such as GIANT, use a two-stage pipeline: first embedding the text attributes of graph nodes then feeding the resulting embeddings into a node classification model. In this paper, we eliminate these two stages and develop an end-to-end node classification model that builds upon GIANT, called End-to-End-GIANT (E2EG). The tandem utilization of a main and an auxiliary classification objectives in our approach results in a more robust model, enabling the BERT backbone to be switched out for a distilled encoder with a 25% - 40% reduction in the number of parameters. Moreover, the model’s end-to-end nature increases ease of use, as it avoids the need of chaining multiple models for node classification. Compared to a GIANT+MLP baseline on the ogbn-arxiv and ogbn-products datasets, E2EG obtains slightly better accuracy in the transductive setting (+0.5%), while reducing model training time by up to 40%. Our model is also applicable in the inductive setting, outperforming GIANT+MLP by up to +2.23%.
Document type Conference contribution
Language English
Published at https://doi.org/10.1109/ICDMW60847.2023.00142
Other links https://github.com/TuAnh23/E2EG
Downloads
Permalink to this page
Back