We present a novel approach to learn representations for sentence-level
semantic similarity using conversational data. Our method trains an
unsupervised model to predict conversational input-response pairs. The
resulting sentence embeddings perform well on the semantic textual similarity
(STS) benchmark and SemEval 2017's Community Question Answering (CQA) question
similarity subtask. Performance is further improved by introducing multitask
training combining the conversational input-response prediction task and a
natural language inference task. Extensive experiments show the proposed model
achieves the best performance among all neural models on the STS benchmark and
is competitive with the state-of-the-art feature engineered and mixed systems
in both tasks.
Description
Learning Semantic Textual Similarity from Conversations
%0 Generic
%1 yang2018learning
%A Yang, Yinfei
%A Yuan, Steve
%A Cer, Daniel
%A Kong, Sheng-yi
%A Constant, Noah
%A Pilar, Petr
%A Ge, Heming
%A Sung, Yun-Hsuan
%A Strope, Brian
%A Kurzweil, Ray
%D 2018
%K matching nlp semantic semantic-measure
%T Learning Semantic Textual Similarity from Conversations
%U http://arxiv.org/abs/1804.07754
%X We present a novel approach to learn representations for sentence-level
semantic similarity using conversational data. Our method trains an
unsupervised model to predict conversational input-response pairs. The
resulting sentence embeddings perform well on the semantic textual similarity
(STS) benchmark and SemEval 2017's Community Question Answering (CQA) question
similarity subtask. Performance is further improved by introducing multitask
training combining the conversational input-response prediction task and a
natural language inference task. Extensive experiments show the proposed model
achieves the best performance among all neural models on the STS benchmark and
is competitive with the state-of-the-art feature engineered and mixed systems
in both tasks.
@misc{yang2018learning,
abstract = {We present a novel approach to learn representations for sentence-level
semantic similarity using conversational data. Our method trains an
unsupervised model to predict conversational input-response pairs. The
resulting sentence embeddings perform well on the semantic textual similarity
(STS) benchmark and SemEval 2017's Community Question Answering (CQA) question
similarity subtask. Performance is further improved by introducing multitask
training combining the conversational input-response prediction task and a
natural language inference task. Extensive experiments show the proposed model
achieves the best performance among all neural models on the STS benchmark and
is competitive with the state-of-the-art feature engineered and mixed systems
in both tasks.},
added-at = {2019-06-16T20:33:29.000+0200},
author = {Yang, Yinfei and Yuan, Steve and Cer, Daniel and Kong, Sheng-yi and Constant, Noah and Pilar, Petr and Ge, Heming and Sung, Yun-Hsuan and Strope, Brian and Kurzweil, Ray},
biburl = {https://www.bibsonomy.org/bibtex/24e96954b74fbafac8004b58345eb4dec/karime},
description = {Learning Semantic Textual Similarity from Conversations},
interhash = {a984ac4f3c61604019efe2fb99698a01},
intrahash = {4e96954b74fbafac8004b58345eb4dec},
keywords = {matching nlp semantic semantic-measure},
note = {cite arxiv:1804.07754Comment: 10 pages, 8 Figures, 6 Tables},
timestamp = {2019-06-16T20:33:29.000+0200},
title = {Learning Semantic Textual Similarity from Conversations},
url = {http://arxiv.org/abs/1804.07754},
year = 2018
}