pith. machine review for the scientific record. sign in

arxiv: 2605.09863 · v1 · submitted 2026-05-11 · 💻 cs.CR · cs.AI· cs.CL· cs.IR· cs.LG

Recognition: no theorem link

Nautilus Compass: Black-box Persona Drift Detection for Production LLM Agents

Chunxiao Wang

Pith reviewed 2026-05-12 04:52 UTC · model grok-4.3

classification 💻 cs.CR cs.AIcs.CLcs.IRcs.LG
keywords persona drift detectionblack-box LLM agentsembedding similarityagent memory layerproduction monitoringdrift detectionclosed API agents
0
0 comments X

The pith

A black-box system detects persona drift in closed-API LLM agents by comparing prompt embeddings to behavioral anchors.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Production LLM coding agents lose track of user constraints over long sessions, repeating flagged errors and forgetting prior agreements. White-box methods that inspect model internals cannot be used on the closed APIs that most users actually run, such as Claude or GPT-4. Nautilus Compass addresses this by working entirely at the prompt-text layer: it embeds raw user prompts and behavioral anchor texts with BGE-m3, then aggregates cosine similarities via weighted top-k mean to signal drift. The design skips LLM calls for fact extraction at index time, ships as a plugin and API server with tamper-evident logging, and is evaluated on real session traces. On held-out data labeled by an independent judge the detector reaches 0.83 ROC AUC while its retrieval layer scores 56.6 percent on LongMemEval-S and 44.4 percent on EverMemBench-Dynamic at low reproduction cost.

Core claim

Nautilus Compass is a black-box persona drift detector and agent memory layer that operates solely on prompt text. It embeds raw conversation turns and computes cosine similarity to a fixed set of behavioral anchor texts, then aggregates the scores by weighted top-k mean using BGE-m3 embeddings. This produces a drift signal without requiring model weights or LLM calls to extract facts. On a held-out test set of real Claude Code session traces labeled by an independent LLM judge, the method reaches ROC AUC 0.83 for drift detection. The same embedding pipeline scores 56.6 percent on LongMemEval-S v0.8 and 44.4 percent on EverMemBench-Dynamic while remaining approximately thirty points below白盒白

What carries the argument

Cosine similarity between raw user prompts and behavioral anchor texts, aggregated by weighted top-k mean on BGE-m3 embeddings, used as the sole drift signal and memory index without fact extraction or model access.

If this is right

  • The detector can be deployed as a Claude Code plugin, MCP server for Cursor and similar tools, CLI, or REST API with a Merkle-chained audit log.
  • The memory layer avoids LLM calls at index time, keeping end-to-end reproduction cost around $3.50.
  • Retrieval performance is bounded by the no-extraction architecture, remaining roughly thirty points below recent white-box leaders.
  • The system tops the four published baselines on EverMemBench-Dynamic while providing system-level A/B evidence across vendors.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar embedding-only checks could be added to other black-box agent frameworks to support longer consistent sessions without API changes.
  • The performance gap to white-box methods indicates a practical trade-off that favors simplicity and cost in production settings.
  • Replacing the LLM judge with direct human labels on drift events would provide a stronger test of whether the similarity signal matches actual user-perceived drift.
  • The approach could be extended to combine prompt similarity with other cheap black-box signals such as response consistency metrics.

Load-bearing premise

That cosine similarity between raw prompts and behavioral anchors, without explicit fact extraction, accurately reflects persona drift and that LLM-judge labels on the test traces constitute reliable ground truth.

What would settle it

A new collection of session traces annotated for drift by human experts in which the method's ROC AUC falls below 0.70 would falsify the detection claim.

Figures

Figures reproduced from arXiv: 2605.09863 by Chunxiao Wang.

Figure 1
Figure 1. Figure 1: Drift detection ROC AUC across the four-step methodology evolution. Each step is a controlled [PITH_FULL_IMAGE:figures/full_fig_p002_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: System architecture. A user-issued prompt enters [PITH_FULL_IMAGE:figures/full_fig_p005_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Drift score distribution for the v0.7.1 final configuration. [PITH_FULL_IMAGE:figures/full_fig_p009_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Reranker MRR lift on LongMemEval-S subset 12 by question type. Hardest question type [PITH_FULL_IMAGE:figures/full_fig_p011_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Per-question-type MRR on LongMemEval-S subset 12. nautilus-compass (m3+rerank) vs. [PITH_FULL_IMAGE:figures/full_fig_p012_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Cross-vendor behavior steering A/B heatmap. Each cell is the [PITH_FULL_IMAGE:figures/full_fig_p016_6.png] view at source ↗
read the original abstract

Production LLM coding agents drift over long sessions: they forget user-specified constraints, slip into mistakes the user already flagged, and confabulate prior agreements. White-box approaches such as persona vectors require model weights and so cannot be applied to closed APIs (Claude, GPT-4) that most users actually interact with. We present Nautilus Compass, a black-box persona drift detector and agent memory layer for production coding agents. The method operates entirely at the prompt-text layer: cosine similarity between user prompts and behavioral anchor texts, aggregated by a weighted top-k mean using BGE-m3 embeddings. Compass is, to our knowledge, the only public agent memory layer (among Mem0, Letta, Cognee, Zep, MemOS, smrti verified May 2026) that does not call an LLM at index time to extract facts or build a graph; raw conversation text is embedded directly. The system ships as a Claude Code plugin, an MCP 2024-11-05 A2A server (Cursor, Cline, Hermes), a CLI, and a REST API on one daemon, with a Merkle-chained audit log for tamper-evident anchor updates. On a held-out test set built from real Claude Code session traces and labeled by an independent LLM judge, Compass reaches ROC AUC 0.83 for drift detection. The embedded retrieval pipeline scores 56.6% on LongMemEval-S v0.8 and 44.4% on EverMemBench-Dynamic (n=500), topping the four published EverMemBench Table 4 baselines. LongMemEval-S 56.6% is ~30 points below recent white-box leaders (90+%); we treat that as the architectural ceiling of the no-extraction design. End-to-end reproduction cost is $3.50 (~14x cheaper than GPT-4o-judged stacks). A paired cross-vendor behavior A/B accompanies these numbers as preliminary system-level evidence. Code, anchors, frozen test data, and audit-log tooling are MIT-licensed at github.com/chunxiaoxx/nautilus-compass.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper presents Nautilus Compass, a black-box persona drift detector and memory layer for production LLM coding agents that operates solely at the prompt-text layer using cosine similarity between raw user prompts and behavioral anchor texts, aggregated by weighted top-k mean on BGE-m3 embeddings. It requires no LLM calls for fact extraction at index time and is deployed as a Claude Code plugin, MCP server, CLI, and REST API with a Merkle-chained audit log. On a held-out test set of real Claude Code traces labeled by an independent LLM judge, the method reports 0.83 ROC AUC for drift detection; the retrieval pipeline scores 56.6% on LongMemEval-S v0.8 and 44.4% on EverMemBench-Dynamic (n=500), outperforming four published baselines on the latter while remaining ~30 points below white-box leaders. Code, anchors, frozen test data, and tooling are released under MIT license.

Significance. If the evaluation holds, the work provides a practical, low-cost ($3.50 reproduction), and reproducible black-box alternative for monitoring persona drift in closed-API agents that dominate production use, addressing a clear gap left by white-box persona-vector methods. The open release of code, anchors, frozen test data, and audit tooling is a notable strength that enables independent verification and extension. The no-extraction design choice is explicitly framed as an architectural trade-off, which aids interpretability of the reported performance gap.

major comments (2)
  1. [Abstract and Results] Abstract and Results section (held-out test set paragraph): The central claim of 0.83 ROC AUC relies on labels from an 'independent LLM judge' on real Claude Code traces, yet no details are supplied on the judge prompt, exact decision criteria for drift (e.g., forgetting constraints, repeating flagged mistakes), calibration, or any human validation of the labels. This is load-bearing because the detector itself uses semantic embeddings without explicit fact extraction; any shared bias between BGE-m3 and the judge model could inflate the AUC by measuring agreement rather than independent drift detection.
  2. [Results] Results section (EverMemBench paragraph): The claim that the pipeline 'tops the four published EverMemBench Table 4 baselines' at 44.4% is presented without describing how those baselines were re-run, whether the identical n=500 subset and evaluation protocol were used, or any statistical significance testing. This undermines the comparative evidence for the no-extraction design.
minor comments (2)
  1. [Abstract] The abstract states 'verified May 2026' for competing systems; this future date should be corrected or clarified as a projection.
  2. [Method] No explicit description of how the behavioral anchors were authored or selected appears in the provided text; adding a short methods paragraph would improve reproducibility.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback, which identifies key areas where additional transparency will strengthen the paper. We address each major comment below and will revise the manuscript accordingly.

read point-by-point responses
  1. Referee: [Abstract and Results] Abstract and Results section (held-out test set paragraph): The central claim of 0.83 ROC AUC relies on labels from an 'independent LLM judge' on real Claude Code traces, yet no details are supplied on the judge prompt, exact decision criteria for drift (e.g., forgetting constraints, repeating flagged mistakes), calibration, or any human validation of the labels. This is load-bearing because the detector itself uses semantic embeddings without explicit fact extraction; any shared bias between BGE-m3 and the judge model could inflate the AUC by measuring agreement rather than independent drift detection.

    Authors: We agree that the labeling methodology requires fuller documentation to support the central claim and to permit assessment of bias risks. In the revised manuscript we will add the complete judge prompt to an appendix, enumerate the exact decision criteria (forgetting user-specified constraints, repetition of previously flagged mistakes, and confabulation of prior agreements), and describe the calibration steps. The judge model (GPT-4o) was deliberately chosen from a different family than BGE-m3; we will expand the discussion of this choice and the residual semantic-bias risk. A full human validation study was not performed; we will state this limitation explicitly. These additions will be incorporated in the Results section and supplementary material. revision: yes

  2. Referee: [Results] Results section (EverMemBench paragraph): The claim that the pipeline 'tops the four published EverMemBench Table 4 baselines' at 44.4% is presented without describing how those baselines were re-run, whether the identical n=500 subset and evaluation protocol were used, or any statistical significance testing. This undermines the comparative evidence for the no-extraction design.

    Authors: We thank the referee for highlighting this gap in the comparative evaluation. In the revised Results section we will document that the four published baselines were re-implemented from their original code, executed on the identical n=500 subset of EverMemBench-Dynamic, and scored under the exact protocol defined in the benchmark. We will also report statistical significance testing (bootstrap confidence intervals and paired tests) for the observed differences. These details will be added to support the claim that the no-extraction pipeline outperforms the listed baselines. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The paper defines a concrete retrieval method (cosine similarity on raw BGE-m3 embeddings of user prompts versus fixed behavioral anchor texts, aggregated by weighted top-k) and reports its empirical performance on held-out Claude Code traces and public benchmarks. No equations, first-principles derivations, or fitted parameters appear in the provided text. The evaluation uses an external LLM judge only for test-set labeling and does not feed those labels back into the detector or any claimed prediction. No self-citations, uniqueness theorems, or ansatzes are invoked to justify the core design. The reported AUC and benchmark scores are therefore measured outcomes rather than quantities that reduce to the method's inputs by construction.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central claim rests on the domain assumption that embedding similarity to anchors proxies persona drift without needing explicit extraction. No new physical entities are postulated; the approach uses a pre-existing embedding model and standard similarity metric.

free parameters (1)
  • top-k and weighting scheme
    The weighted top-k mean aggregation likely involves tunable parameters for k and weights, though exact values are not specified in the abstract.
axioms (1)
  • domain assumption Cosine similarity on BGE-m3 embeddings of raw text captures behavioral consistency relevant to persona drift
    Invoked as the core detection mechanism operating at the prompt-text layer.

pith-pipeline@v0.9.0 · 5696 in / 1486 out tokens · 78757 ms · 2026-05-12T04:52:08.734638+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

11 extracted references · 11 canonical work pages · 7 internal anchors

  1. [2]

    A-MEM: Agentic Memory for LLM Agents

    arXiv preprint arXiv:2502.12110. Anonymous. Textdefendr: Empirical robustness of black-box llm safety filters,

  2. [4]

    M3-Embedding: Multi-Linguality, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation

    Jianlv Chen, Shitao Xiao, Peitian Zhang, Kun Luo, Defu Lian, and Zheng Liu. M3-embedding: Multi- linguality, multi-functionality, multi-granularity text embeddings through self-knowledge distillation. arXiv preprint arXiv:2402.03216,

  3. [6]

    Persona Vectors: Monitoring and Controlling Character Traits in Language Models

    URL https://arxiv.org/abs/2507.21509. DPT-Agent Authors. DPT-Agent: Distillation path-triggered agent for memory-augmented reasoning. https://arxiv.org/abs/2502.11882,

  4. [7]

    Laura Hanu and Unitary

    arXiv preprint arXiv:2502.11882. Laura Hanu and Unitary. Detoxify: Toxic comment classification with multilingual transformers.https: //github.com/unitaryai/detoxify,

  5. [8]

    task-updates

    Chris Yuhao Liu, Liang Zeng, Jiacai Liu, et al. Skywork-reward: Reward modeling at scale for practical use.arXiv preprint arXiv:2410.18451,

  6. [9]

    Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory

    arXiv preprint arXiv:2504.19413; accessed 2026-05-07. 18 OpenAI. Moderation api·classification of unsafe content. https://platform.openai.com/docs/ guides/moderation,

  7. [10]

    MemGPT: Towards LLMs as Operating Systems

    Charles Packer, Sarah Wooders, Kevin Lin, Vivian Fang, Shishir G Patil, Ion Stoica, and Joseph E Gonzalez. Memgpt: Towards llms as operating systems.arXiv preprint arXiv:2310.08560,

  8. [11]

    Zep: A Temporal Knowledge Graph Architecture for Agent Memory

    Preston Rasmussen, Pavlo Paliychuk, Travis Beauvais, Jack Ryan, and Daniel Chalef. Zep: A temporal knowledge graph architecture for agent memory.arXiv preprint arXiv:2501.13956,

  9. [12]

    Sentence-bert: Sentence embeddings using siamese bert-networks

    Nils Reimers and Iryna Gurevych. Sentence-bert: Sentence embeddings using siamese bert-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing,

  10. [13]

    Helpsteer 2-preference: Complementing ratings with preferences.arXiv preprint arXiv:2410.01257,

    Zhilin Wang, Alexander Bukharin, Olivier Delalleau, et al. Helpsteer 2-preference: Complementing ratings with preferences.arXiv preprint arXiv:2410.01257,

  11. [14]

    LongMemEval: Benchmarking Chat Assistants on Long-Term Interactive Memory

    Di Wu, Hongwei Wang, Wenhao Yu, Yuwei Zhang, Kai-Wei Chang, and Dong Yu. Longmemeval: Benchmarking chat assistants on long-term interactive memory.arXiv preprint arXiv:2410.10813,