pith. machine review for the scientific record. sign in

arxiv: 2604.15728 · v1 · submitted 2026-04-17 · 💻 cs.CR · cs.AI

Recognition: unknown

Privacy-Preserving LLMs Routing

Qian Lou, Reza Shirkavand, Shangqian Gao, Xidong Wu, Yukuan Zhang, Yuqiong Ji

Authors on Pith no claims yet

Pith reviewed 2026-05-10 08:31 UTC · model grok-4.3

classification 💻 cs.CR cs.AI
keywords pprouteroutingmodelprivacy-preservingalgorithmcommunicationcomputationencoder
0
0 comments X

The pith

PPRoute achieves plaintext-level LLM routing quality with MPC-based privacy and a 20x speedup over naive encrypted implementations via MPC-friendly encoders, multi-step training, and O(1) communication Top-k search.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Large language models are expensive to run, so companies route queries to different models based on cost and quality. But the router sees the user's prompt, creating a privacy problem. PPRoute solves this by using secure multi-party computation so the router never sees the actual data. To avoid the usual slowdown from encryption, the authors replace normal operations with MPC-friendly versions, train the router in stages that work in the encrypted space, and replace slow secure sorting with a new unsorted Top-k method that needs almost no extra communication. Tests on several datasets show the private version matches the accuracy of the non-private router while running about twenty times faster than a basic encrypted version.

Core claim

Across different datasets, PPRoute achieves the performance of plaintext counterparts, while achieving approximately a 20× speedup over naïve MPC implementations.

Load-bearing premise

The multi-step training algorithm and MPC-friendly operations preserve routing quality without introducing privacy leaks or accuracy drops that would make the system unusable in real deployments.

Figures

Figures reproduced from arXiv: 2604.15728 by Qian Lou, Reza Shirkavand, Shangqian Gao, Xidong Wu, Yukuan Zhang, Yuqiong Ji.

Figure 1
Figure 1. Figure 1: Overview of Privacy-Preserving LLMs Routing (PPRoute) Training. 1) Train the MLP layer in the teacher model. 2) The student model uses the MPC-friendly operations in the approximated encoder, and model training follows Equation (2). 3) Train the MLP layers only and encrypt the model to serve under the MPC. (mod 2l ) without communication (Wagh et al., 2019). Beaver (1991); Juvekar et al. (2018) address the… view at source ↗
Figure 2
Figure 2. Figure 2: Overview of Privacy-Preserving LLMs Routing (PPRoute) Inference. 1) Generate query embedding; 2) Use Unsorted-topk to do retrieval or clustering. GeLU approximation. The Gaussian error function is com￾putationally expensive and suffers from reduced numerical accuracy under MPC. Since ReLU(x) is simple, efficient, and widely adopted in machine learning models, PPRoute uses ReLU(x) to approximate GeLU activa… view at source ↗
Figure 3
Figure 3. Figure 3: The Unsorted Top-k Algorithm Workflow MPC is usually communication complexity, the indepen￾dence of each matrix entry allows parallel execution and the size of the model pool is limited. Consequently, com￾munication complexity is bounded by the constant depth of a single secure comparison, O(1), effectively mitigat￾ing the latency overhead typically associated with iterative comparison protocols. Following… view at source ↗
Figure 4
Figure 4. Figure 4: Accuracy–cost/size deferral curves on three datasets performance across various datasets compared to UniRoute. This highlights that PPRoute can be widely applied to vari￾ous embedding-based LLM routing and reduce privacy risk. 5. Conclusion Although LLM routing is essential for balancing model performance and cost-efficiency, it introduces significant privacy risks that have remained largely unaddressed. T… view at source ↗
read the original abstract

Large language model (LLM) routing has emerged as a critical strategy to balance model performance and cost-efficiency by dynamically selecting services from various model providers. However, LLM routing adds an intermediate layer between users and LLMs, creating new privacy risks to user data. These privacy risks have not been systematically studied. Although cryptographic techniques such as Secure Multi-Party Computation (MPC) enable privacy-preserving computation, their protocol design and implementation remain under-explored, and na\"ive implementations typically incur prohibitive computational overhead. To address this, we propose a privacy-preserving LLM routing framework (PPRoute). PPRoute includes multiple strategies to speed up encoder inference and nearest neighbor search under the MPC and maintain the quality of LLM routing. First, PPRoute uses MPC-friendly operations to boost the encoder inference. Second, PPRoute uses a multiple-step model training algorithm to maintain routing quality despite the constraints of the encrypted domain. Third, PPRoute proposes an unsorted Top-k algorithm with $O(1)$ communication complexity for secure sorting in model search, significantly reducing communication latency. Across different datasets, PPRoute achieves the performance of plaintext counterparts, while achieving approximately a 20$\times$ speedup over na\"ive MPC implementations.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The paper describes an empirical system (PPRoute) built from MPC-friendly encoder operations, a multi-step training procedure, and an unsorted Top-k search with claimed O(1) communication. All central performance claims are presented as measured outcomes on datasets rather than as predictions derived from fitted parameters or self-referential equations. No derivation chain reduces a result to its own inputs by construction, and no load-bearing uniqueness theorem or ansatz is imported via self-citation. The framework is therefore self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

The abstract supplies no explicit free parameters, axioms, or invented entities. The framework implicitly assumes that MPC can be made efficient enough for real-time routing and that staged training can recover quality lost to encryption constraints.

pith-pipeline@v0.9.0 · 5525 in / 1127 out tokens · 45884 ms · 2026-05-10T08:31:28.660877+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

21 extracted references · 17 canonical work pages · 1 internal anchor

  1. [1]

    Automix: Automatically mixing language models

    Aggarwal, P., Madaan, A., Anand, A., Potharaju, S. P., Mishra, S., Zhou, P., Gupta, A., Rajagopal, D., Kappa- ganthu, K., Yang, Y ., et al. Automix: Automatically mix- ing language models.arXiv preprint arXiv:2310.12963,

  2. [2]

    Batcher, K. E. Sorting networks and their applications. In Proceedings of the April 30–May 2, 1968, spring joint computer conference, pp. 307–314,

  3. [3]

    FrugalGPT: How to Use Large Language Models While Reducing Cost and Improving Performance

    Chen, L., Zaharia, M., and Zou, J. Frugalgpt: How to use large language models while reducing cost and improving performance.arXiv preprint arXiv:2305.05176,

  4. [4]

    Hybrid llm: Cost-efficient and quality- aware query routing.arXiv preprint arXiv:2404.14618, 2024

    Ding, D., Mallick, A., Wang, C., Sim, R., Mukherjee, S., Ruhle, V ., Lakshmanan, L. V ., and Awadallah, A. H. Hy- brid llm: Cost-efficient and quality-aware query routing. arXiv preprint arXiv:2404.14618,

  5. [5]

    arXiv preprint arXiv:2410.03834 , year=

    Feng, T., Shen, Y ., and You, J. Graphrouter: A graph-based router for llm selections.arXiv preprint arXiv:2410.03834,

  6. [6]

    Huang, L., Yu, W., Ma, W., Zhong, W., Feng, Z., Wang, H., Chen, Q., Peng, W., Feng, X., Qin, B., and Liu, T

    Hu, Q. J., Bieker, J., Li, X., Jiang, N., Keigwin, B., Ran- ganath, G., Keutzer, K., and Upadhyay, S. K. Router- bench: A benchmark for multi-llm routing system.arXiv preprint arXiv:2403.12031,

  7. [7]

    arXiv preprint arXiv:2510.19506 , year=

    Huang, C., Shi, T., Zhu, Y ., Chen, R., and Quan, X. Looka- head routing for large language models.arXiv preprint arXiv:2510.19506,

  8. [8]

    Jiang, D., Ren, X., and Lin, B. Y . Llm-blender: Ensembling large language models with pairwise ranking and genera- tive fusion.arXiv preprint arXiv:2306.02561,

  9. [9]

    Universal Model Routing for Efficient LLM Inference.arXiv preprint arXiv:2502.08773, 2025

    Jitkrittum, W., Narasimhan, H., Rawat, A. S., Juneja, J., Wang, C., Wang, Z., Go, A., Lee, C.-Y ., Shenoy, P., Pani- grahy, R., et al. Universal model routing for efficient llm inference.arXiv preprint arXiv:2502.08773,

  10. [10]

    Lu, K., Yuan, H., Lin, R., Lin, J., Yuan, Z., Zhou, C., and Zhou, J

    URL https: //openreview.net/forum?id=1VqxIgyQlp. Lu, K., Yuan, H., Lin, R., Lin, J., Yuan, Z., Zhou, C., and Zhou, J. Routing to the expert: Efficient reward-guided ensemble of large language models. InProceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (V olume 1: ...

  11. [11]

    SecFormer: Fast and accurate privacy-preserving inference for transformer models via SMPC

    9 Submission and Formatting Instructions for ICML 2026 Luo, J., Zhang, Y ., Zhang, Z., Zhang, J., Mu, X., Wang, H., Yu, Y ., and Xu, Z. SecFormer: Fast and accurate privacy-preserving inference for transformer models via SMPC. In Ku, L.-W., Martins, A., and Srikumar, V . (eds.),Findings of the Association for Computational Linguistics: ACL 2024, pp. 13333...

  12. [12]

    Rethinking tabular data understanding with large language models

    Association for Computational Linguistics. doi: 10.18653/v1/2024. findings-acl.790. URL https://aclanthology. org/2024.findings-acl.790/. Malkov, Y . A. and Yashunin, D. A. Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs.IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(4):824– 836,

  13. [13]

    Malkov and D

    doi: 10.1109/TPAMI.2018.2889473. Mohassel, P. and Zhang, Y . Secureml: A system for scalable privacy-preserving machine learning. In2017 IEEE sym- posium on security and privacy (SP), pp. 19–38. IEEE,

  14. [14]

    Cost- aware contrastive routing for llms.arXiv preprint arXiv:2508.12491,

    Shirkavand, R., Gao, S., Yu, P., and Huang, H. Cost- aware contrastive routing for llms.arXiv preprint arXiv:2508.12491,

  15. [15]

    Large language model routing with benchmark datasets.arXiv preprint arXiv:2309.15789, 2023

    Shnitzer, T., Ou, A., Silva, M., Soule, K., Sun, Y ., Solomon, J., Thompson, N., and Yurochkin, M. Large language model routing with benchmark datasets.arXiv preprint arXiv:2309.15789,

  16. [16]

    D., Jin, H., Yao, Y ., Zhang, J., Zhang, T., Avestimehr, S., and He, C

    Stripelis, D., Xu, Z., Hu, Z., Shah, A. D., Jin, H., Yao, Y ., Zhang, J., Zhang, T., Avestimehr, S., and He, C. Ten- soropera router: A multi-model router for efficient llm inference. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: In- dustry Track, pp. 452–462,

  17. [17]

    Causal llm routing: End-to-end regret minimization from observational data.arXiv preprint arXiv:2505.16037, 2025

    Tsiourvas, A., Sun, W., and Perakis, G. Causal llm routing: End-to-end regret minimization from observational data. arXiv preprint arXiv:2505.16037,

  18. [18]

    When to reason: Semantic router for vllm

    Wang, C., Liu, X., Liu, Y ., Zhu, Y ., Mo, X., Jiang, J., and Chen, H. When to reason: Semantic router for vllm. arXiv preprint arXiv:2510.08731,

  19. [19]

    arXiv preprint arXiv:2310.03094 , year=

    Yue, M., Zhao, J., Zhang, M., Du, L., and Yao, Z. Large language model cascades with mixture of thoughts rep- resentations for cost-efficient reasoning.arXiv preprint arXiv:2310.03094,

  20. [20]

    Mpcache: Mpc-friendly kv cache eviction for efficient private large language model inference.arXiv preprint arXiv:2501.06807,

    Zeng, W., Dong, Y ., Zhou, J., Ma, J., Tan, J., Wang, R., and Li, M. Mpcache: Mpc-friendly kv cache eviction for efficient private large language model inference.arXiv preprint arXiv:2501.06807,

  21. [21]

    Office Products

    Zhuang, R., Wu, T., Wen, Z., Li, A., Jiao, J., and Ramchan- dran, K. Embedllm: Learning compact representations of large language models.arXiv preprint arXiv:2410.02223,