pith. machine review for the scientific record. sign in

arxiv: 1705.00652 · v1 · submitted 2017-05-01 · 💻 cs.CL

Recognition: unknown

Efficient Natural Language Response Suggestion for Smart Reply

Authors on Pith no claims yet
classification 💻 cs.CL
keywords responseefficientlanguagemethodnaturaloptimizedsuggestionachieves
0
0 comments X
read the original abstract

This paper presents a computationally efficient machine-learned method for natural language response suggestion. Feed-forward neural networks using n-gram embedding features encode messages into vectors which are optimized to give message-response pairs a high dot-product value. An optimized search finds response suggestions. The method is evaluated in a large-scale commercial e-mail application, Inbox by Gmail. Compared to a sequence-to-sequence approach, the new system achieves the same quality at a small fraction of the computational requirements and latency.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 5 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. SimCSE: Simple Contrastive Learning of Sentence Embeddings

    cs.CL 2021-04 conditional novelty 8.0

    SimCSE achieves 76.3% unsupervised and 81.6% supervised Spearman's correlation on STS tasks with BERT-base, improving prior best results by 4.2% and 2.2% via simple contrastive learning.

  2. ProtSent: Protein Sentence Transformers

    cs.LG 2026-05 unverdicted novelty 7.0

    Contrastive fine-tuning of protein language models on Pfam, structural, interaction, and mutational datasets produces embeddings that improve kNN performance on 15-16 of 23 downstream tasks including remote homology d...

  3. ORPHEAS: A Cross-Lingual Greek-English Embedding Model for Retrieval-Augmented Generation

    cs.CL 2026-04 unverdicted novelty 6.0

    ORPHEAS, a Greek-English embedding model created with knowledge graph fine-tuning, outperforms state-of-the-art multilingual models on monolingual and cross-lingual retrieval benchmarks.

  4. Domain-Adapted Retrieval for In-Context Annotation of Pedagogical Dialogue Acts

    cs.CL 2026-04 unverdicted novelty 5.0

    Domain-adapted utterance-level retrieval raises Cohen's kappa for tutoring dialogue act annotation to 0.526-0.580 on TalkMoves and 0.659-0.743 on Eedi, beating no-retrieval baselines by large margins across three LLMs.

  5. Unified Supervision for Walmart's Sponsored Search Retrieval via Joint Semantic Relevance and Behavioral Engagement Modeling

    cs.IR 2026-04 unverdicted novelty 4.0

    A hybrid supervision method for bi-encoder retrievers combines graded relevance from teacher models, production retrieval priors, and selective engagement to improve relevance and NDCG over Walmart's current sponsored...