Pre-train, Interact, Fine-tune: A Novel Interaction Representation for Text Classification

Open Access
Authors
Publication date 11-2020
Journal Information Processing & Management
Article number 102215
Volume | Issue number 57 | 6
Number of pages 15
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Text representation can aid machines in text understanding. Previous work on text representation often focuses on the so-called forward implication, i.e., preceding words are taken as the context of later words for creating representations, effective it is, yet ignoring the fact that the semantics of a text segment is a product of the mutual implication of words in the text: later words contribute to the meaning of preceding words. To bridge this gap, we introduce the concept of interaction and propose a two-perspective interaction representation, in which it encapsulates a local and a global interaction representation. Here, a local interaction representation is one that interacts among words with parent-children relationships on the syntactic trees whereas a global interaction interpretation is one that interacts among all the words in a sentence. We combine these two interaction representations to develop a Hybrid Interaction Representation (HIR).
Inspired by existing feature-based and fine-tuning-based pretrain-finetuning approaches to language models, we integrate the merits of feature-based and fine-tuning-based methods to propose the Pre-train, Interact, Fine-tune (PIF) architecture.
We evaluate our proposed models on five widely-used datasets for text classification tasks. It turns out that our ensemble method, HIRP, outperforms state-of-the-art baselines with improvements ranging from 2.03% to 3.15% in terms of error rate. In addition, we find that, the improvements of PIF against most state-of-the-art methods is not affected by increasing of the text length.

Document type Article
Language English
Related publication Pre-train, Interact, Fine-tune: A Novel Interaction Representation for Text Classification
Published at https://doi.org/10.1016/j.ipm.2020.102215
Published at https://arxiv.org/abs/1909.11824
Downloads
Pre-train, Interact, Fine-tune arxiv (Submitted manuscript)
Permalink to this page
Back