pith. machine review for the scientific record. sign in

arxiv: 1608.04207 · v3 · submitted 2016-08-15 · 💻 cs.CL

Recognition: unknown

Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks

Authors on Pith no claims yet
classification 💻 cs.CL
keywords sentencerepresentationspredictiontasksencodedlengthvectorsword
0
0 comments X
read the original abstract

There is a lot of research interest in encoding variable length sentences into fixed length vectors, in a way that preserves the sentence meanings. Two common methods include representations based on averaging word vectors, and representations based on the hidden states of recurrent neural networks such as LSTMs. The sentence vectors are used as features for subsequent machine learning tasks or for pre-training in the context of deep learning. However, not much is known about the properties that are encoded in these sentence representations and about the language information they capture. We propose a framework that facilitates better understanding of the encoded representations. We define prediction tasks around isolated aspects of sentence structure (namely sentence length, word content, and word order), and score representations by the ability to train a classifier to solve each prediction task when using the representation as input. We demonstrate the potential contribution of the approach by analyzing different sentence representation mechanisms. The analysis sheds light on the relative strengths of different sentence embedding methods with respect to these low level prediction tasks, and on the effect of the encoded vector's dimensionality on the resulting representations.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 4 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Pointer Sentinel Mixture Models

    cs.CL 2016-09 conditional novelty 7.0

    Pointer sentinel-LSTM mixes context copying with softmax prediction to reach 70.9 perplexity on Penn Treebank using fewer parameters than standard LSTMs.

  2. Finding Meaning in Embeddings: Concept Separation Curves

    cs.CL 2026-04 unverdicted novelty 6.0

    Concept Separation Curves provide a classifier-independent method to visualize and quantify how sentence embeddings distinguish conceptual meaning from syntactic variations across languages and domains.

  3. Towards Generalizable Representations of Mathematical Strategies

    cs.CY 2026-04 unverdicted novelty 6.0

    Transition embeddings from pretrained algebraic encoders combined with SimCSE produce problem-invariant representations of student solution strategies that encode meaningful information and correlate with short- and l...

  4. Probing Classifiers: Promises, Shortcomings, and Advances

    cs.CL 2021-02 unverdicted novelty 3.0

    Probing classifiers are a common but limited method for analyzing linguistic knowledge in neural NLP models, and this review outlines their promises, methodological shortcomings, and recent advances.