Word n-gram attention models for sentence similarity and inference
Ver/
Fecha
2019-04-22Metadatos
Mostrar el registro completo del ítem
Expert Systems with Applications 132 : 1-11 (2019)
Resumen
Semantic Textual Similarity and Natural Language Inference are two popular natural language understanding tasks used to benchmark sentence representation models where two sentences are paired. In such tasks sentences are represented as bag of words, sequences, trees or convolutions, but the attention model is based on word pairs. In this article we introduce the use of word n-grams in the attention model. Our results on five datasets show an error reduction of up to 41% with respect to the word-based attention model. The improvements are especially relevant with low data regimes and, in the case of natural language inference, on the recently released hard subset of Natural Language Inference datasets.