Abstract

Many modern NLP systems rely on word embeddings, previously trained in an\nunsupervised manner on large corpora, as base features. Efforts to obtain\nembeddings for larger chunks of text, such as sentences, have however not been\nso successful. Several attempts at learning unsupervised representations of\nsentences have not reached satisfactory enough performance to be widely\nadopted. In this paper, we show how universal sentence representations trained\nusing the supervised data of the Stanford Natural Language Inference datasets\ncan consistently outperform unsupervised methods like SkipThought vectors on a\nwide range of transfer tasks. Much like how computer vision uses ImageNet to\nobtain features, which can then be transferred to other tasks, our work tends\nto indicate the suitability of natural language inference for transfer learning\nto other NLP tasks. Our encoder is publicly available.\n

Keywords

Computer scienceArtificial intelligenceNatural language processingInferenceSentenceUnsupervised learningTransfer of learningNatural languageWord (group theory)Machine learningLinguistics

Affiliated Institutions

Related Publications

Publication Info

Year
2017
Type
article
Citations
2038
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2038
OpenAlex

Cite This

Alexis Conneau, Douwe Kiela, Holger Schwenk et al. (2017). Supervised Learning of Universal Sentence Representations from Natural\n Language Inference Data. arXiv (Cornell University) . https://doi.org/10.48550/arxiv.1705.02364

Identifiers

DOI
10.48550/arxiv.1705.02364