Wasserstein barycenter model ensembling
Pierre Dognin, Igor Melnyk, et al.
ICLR 2019
This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention. Instead of using a vector, we use a 2-D matrix to represent the embedding, with each row of the matrix attending on a different part of the sentence. We also propose a self-attention mechanism and a special regularization term for the model. As a side effect, the embedding comes with an easy way of visualizing what specific parts of the sentence are encoded into the embedding. We evaluate our model on 3 different tasks: author profiling, sentiment classification and textual entailment. Results show that our model yields a significant performance gain compared to other sentence embedding methods in all of the 3 tasks.
Pierre Dognin, Igor Melnyk, et al.
ICLR 2019
Mo Yu, Wenpeng Yin, et al.
ACL 2017
Ramesh Nallapati, Bowen Zhou, et al.
CoNLL 2016
Cicero dos Santos, Igor Melnyk, et al.
ACL 2018