pith. machine review for the scientific record. sign in

hub

BERT : Pre-training of deep bidirectional transformers for language understanding

98 Pith papers cite this work, alongside 6,639 external citations. Polarity classification is still indexing.

98 Pith papers citing it
6,639 external citations · Crossref

hub tools

citation-role summary

background 3

citation-polarity summary

claims ledger

  • background The retrieval system only manages to fetch informationabout Fleming's professional achievements in the discoveryof penicillin. However, the document does not provide informa-tion about his educational background, thus the model generates ahallucinatory answer. inappropriately activated, blindly retrieving inaccurate information and consequently leading to an undesirable response. Consequently, several studies [75, 204, 228, 378] have proposed to make a shift from passive retrieval to adaptive re

co-cited works

roles

background 3

polarities

background 3

representative citing papers

How Language Models Process Negation

cs.CL · 2026-05-04 · unverdicted · novelty 7.0

LLMs implement both attention-based suppression and constructive representations for negation, with construction dominant, despite poor accuracy from late-layer attention shortcuts.

Moshi: a speech-text foundation model for real-time dialogue

eess.AS · 2024-09-17 · accept · novelty 7.0

Moshi is the first real-time full-duplex spoken large language model that casts dialogue as speech-to-speech generation using parallel audio streams and an inner monologue of time-aligned text tokens.

citing papers explorer

Showing 50 of 98 citing papers.