Skip to main navigation Skip to search Skip to main content

SANTM: efficient self-attention-driven network for text matching

  • Prayag Tiwari
  • , Amit Kumar Jaiswal
  • , Sahil Garg
  • , Ilsun You
  • University of Padua
  • Aalto University
  • University of Leeds
  • École de technologie supérieure
  • Soonchunhyang University

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)

Abstract

Self-attention mechanisms have recently been embraced for a broad range of text-matching applications. Self-attention model takes only one sentence as an input with no extra information, i.e., one can utilize the final hidden state or pooling. However, text-matching problems can be interpreted either in symmetrical or asymmetrical scopes. For instance, paraphrase detection is an asymmetrical task, while textual entailment classification and question-answer matching are considered asymmetrical tasks. In this article, we leverage attractive properties of self-attention mechanism and proposes an attention-based network that incorporates three key components for inter-sequence attention: global pointwise features, preceding attentive features, and contextual features while updating the rest of the components. Our model follows evaluation on two benchmark datasets cover tasks of textual entailment and question-answer matching. The proposed efficient Self-attention-driven Network for Text Matching outperforms the state of the art on the Stanford Natural Language Inference and WikiQA datasets with much fewer parameters.

Original languageEnglish
Article number55
JournalACM Transactions on Internet Technology
Volume22
Issue number3
DOIs
Publication statusPublished - 29 Nov 2021

Keywords

  • attention mechanism
  • deep learning
  • Text matching

ASJC Scopus subject areas

  • Computer Networks and Communications

Fingerprint

Dive into the research topics of 'SANTM: efficient self-attention-driven network for text matching'. Together they form a unique fingerprint.

Cite this