pith. machine review for the scientific record. sign in

arxiv: 2604.15609 · v1 · submitted 2026-04-17 · 💻 cs.LG · cs.CV

Recognition: unknown

Adapting in the Dark: Efficient and Stable Test-Time Adaptation for Black-Box Models

Chengyi Cai, Feng Liu, Jihun Hamm, Shuaicheng Niu, Yunbei Zhang

Authors on Pith no claims yet

Pith reviewed 2026-05-10 08:50 UTC · model grok-4.3

classification 💻 cs.LG cs.CV
keywords adaptationbetablack-boxefficienttest-timeachievesenablesinference
0
0 comments X

The pith

BETA adapts black-box models at test time using a local steering model and regularization techniques to achieve accuracy improvements without additional API queries or high latency.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

In machine learning, models often face new data that differs from their training data, leading to performance drops. Test-time adaptation tries to adjust the model on the fly to handle this. For black-box models behind commercial APIs, direct changes are impossible and extra queries cost money and time. BETA adds a small, controllable local model that acts as a steering guide to create usable gradients for adaptation. It pairs this with techniques to harmonize predictions, enforce consistency across views, and filter based on prompt learning ideas. The result is adaptation that stays stable in unsupervised settings and requires no more API calls than normal inference. Tests on ImageNet-C show gains over existing methods like TENT and TPT for vision transformers and CLIP models. On a real commercial API, it matches the accuracy of more expensive zeroth-order optimization but at far lower query cost while running in real time.

Core claim

BETA achieves a +7.1% accuracy gain on ViT-B/16 and +3.4% on CLIP on ImageNet-C, surpassing strong white-box and gray-box methods including TENT and TPT; on a commercial API it achieves comparable performance to ZOO at 250x lower cost while maintaining real-time inference speed.

Load-bearing premise

The lightweight local white-box steering model combined with prediction harmonization, consistency regularization, and prompt learning-oriented filtering creates a tractable and stable gradient pathway for unsupervised black-box TTA without requiring additional API calls.

Figures

Figures reproduced from arXiv: 2604.15609 by Chengyi Cai, Feng Liu, Jihun Hamm, Shuaicheng Niu, Yunbei Zhang.

Figure 1
Figure 1. Figure 1: The black-box Test-time Adaptation setting studied in this work. From the client’s perspective, the goal is to adapt a powerful server-side API model to a target distribution (e.g., corrupted or domain-shifted images) without any internal access. The client can only send raw input images and receive softmax probability vectors in return. Unlike white-box TTA, no gradients, parameters, intermediate features… view at source ↗
Figure 2
Figure 2. Figure 2: Comparison of black-box test-time adaptation strategies. (a) Output Refinement (LAME) is limited to post-processing predictions, while (b) ZOO-based Input Prompt Learning requires multiple expensive API calls for prompt optimization. In contrast, (c) BETA achieves efficient single-query adaptation by leveraging a lightweight steering model with prediction harmonization to create a tractable gradient pathwa… view at source ↗
Figure 3
Figure 3. Figure 3: We analyze the trade-off between Objective Relevance (alignment with the true target gradient) and Optimization Ef￾fectiveness (alignment with the practical steering gradient) as a function of α. The intersection of these opposing curves identifies the optimal range (e.g., α ∈ [0.3, 0.5]) where the objective is si￾multaneously relevant to the target and tractable for optimization. The curves are plotted ba… view at source ↗
Figure 4
Figure 4. Figure 4: (a) Five independent runs using solely Eqn. (1) on ImageNet-C (Contrast, level 5) with ViT-B/16 as the target model, showing either performance collapse or failure to improve. (b) Performance vs. API budget on the real-world Clarifai API. clean (x) and prompted (x ′ ) images: Lconsist(x, x′ ) := DKL(pS(x)∥pS(x ′ )) = PK k=1 p k S (x) log p k S (x) p k S (x′) . (5) Final Objective and Joint Optimization. BE… view at source ↗
Figure 5
Figure 5. Figure 5: Sensitivity analysis of BETA’s hyperparameters, showing stable performance across fusion weight α in Eq. 1, regularization weight λ in Eq. 7, entropy margin ϵ in Eq. 4, and prompt size. VLM methods, including TPT and DynaPrompt, as well as gray-box methods such as TCA. This consistent success across datasets and model types demonstrates that BETA is a general framework for black-box adaptation. More result… view at source ↗
Figure 6
Figure 6. Figure 6: Online Batch Accuracy on ImageNet-C Contrast domain. B.10. Robustness to Batch Size In practical online deployment, the number of samples available for adaptation at any given time step can vary significantly. To assess BETA’s sensitivity to this factor, we evaluated its performance on ImageNet-C (ViT-B/16) using batch sizes ranging from 4 to 128. As shown in [PITH_FULL_IMAGE:figures/full_fig_p021_6.png] view at source ↗
read the original abstract

Test-Time Adaptation (TTA) for black-box models accessible only via APIs remains a largely unexplored challenge. Existing approaches such as post-hoc output refinement offer limited adaptive capacity, while Zeroth-Order Optimization (ZOO) enables input-space adaptation but faces high query costs and optimization challenges in the unsupervised TTA setting. We introduce BETA (Black-box Efficient Test-time Adaptation), a framework that addresses these limitations by employing a lightweight, local white-box steering model to create a tractable gradient pathway. Through a prediction harmonization technique combined with consistency regularization and prompt learning-oriented filtering, BETA enables stable adaptation with no additional API calls and negligible latency beyond standard inference. On ImageNet-C, BETA achieves a +7.1% accuracy gain on ViT-B/16 and +3.4% on CLIP, surpassing strong white-box and gray-box methods including TENT and TPT. On a commercial API, BETA achieves comparable performance to ZOO at 250x lower cost while maintaining real-time inference speed, establishing it as a practical and efficient solution for real-world black-box TTA.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The BETA framework introduces independent components—a lightweight local white-box steering model, prediction harmonization, consistency regularization, and prompt-oriented filtering—to enable black-box TTA without extra API calls. These are not shown to reduce by construction to prior fitted quantities, self-referential equations, or load-bearing self-citations; the central claim rests on the new gradient pathway and is supported by external empirical benchmarks (ImageNet-C gains, commercial API comparison) rather than tautological renaming or ansatz smuggling. The derivation chain remains self-contained against the stated inputs.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract describes an empirical engineering method with no explicit mathematical derivations, free parameters, axioms, or newly postulated entities.

pith-pipeline@v0.9.0 · 5508 in / 1118 out tokens · 51295 ms · 2026-05-10T08:50:48.738613+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

9 extracted references · 4 canonical work pages · 3 internal anchors

  1. [1]

    Benchmarking Neural Network Robustness to Common Corruptions and Perturbations

    URL https://api.semanticscholar. org/CorpusID:246015836. Chen, G., Niu, S., Chen, D., Yang, J., Zhang, Z., Tan, M., Wu, P., and Shen, Z. Zerosiam: An efficient asymmetry for test-time entropy optimization without collapse. In The Fourteenth International Conference on Learning Representations, 2026. URL https://openreview. net/forum?id=x6jHZYhnhL. Dosovit...

  2. [2]

    cc/paper_files/paper/2021/file/ 1415fe9fea0fa1e45dddcff5682239a0-Paper

    URL https://proceedings.neurips. cc/paper_files/paper/2021/file/ 1415fe9fea0fa1e45dddcff5682239a0-Paper. pdf. Karmanov, A., Guan, D., Lu, S., El Saddik, A., and Xing, E. Efficient test-time adaptation of vision-language mod- els.The IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024. Kawahara, J., Daneshvar, S., Argenziano, G., and Hamar...

  3. [3]

    and Varshney, Pramod K

    URL https://openreview.net/forum? id=V3zobHnS61. Liu, S., Kailkhura, B., Chen, P.-Y ., Ting, P., Chang, S., and Amini, L. Zeroth-order stochastic variance reduction for nonconvex optimization.Advances in neural information processing systems, 31, 2018. Liu, S., Chen, P.-Y ., Kailkhura, B., Zhang, G., Hero III, A. O., and Varshney, P. K. A primer on zeroth...

  4. [4]

    Gemini: A Family of Highly Capable Multimodal Models

    URL https://openreview.net/forum? id=qz1Vx1v9iK. Oh, C., Hwang, H., Lee, H.-y., Lim, Y ., Jung, G., Jung, J., Choi, H., and Song, K. Blackvip: Black-box visual prompting for robust transfer learning. InProceedings of the IEEE/CVF Conference on Computer Vision and Pat- tern Recognition (CVPR), pp. 24224–24235, June 2023. Ouali, Y ., Bulat, A., Martinez, B....

  5. [5]

    BiomedCLIP: a multimodal biomedical foundation model pretrained from fifteen million scientific image-text pairs

    URL https://openreview.net/forum? id=UwtXDttgsE. Xiao, Z., Yan, S., Hong, J., Cai, J., Jiang, X., Hu, Y ., Shen, J., Wang, Q., and Snoek, C. G. Dynaprompt: Dynamic test-time prompt tuning. InThe Thirteenth International Conference on Learning Representations, 2025. Zhang, C., Stepputtis, S., Sycara, K. P., and Xie, Y . Dual prototype evolving for test-tim...

  6. [6]

    Zhou, L., Ye, M., Li, S., Li, N., Zhu, X., Deng, L., Liu, H., and Lei, Z

    URL https://openreview.net/forum? id=kIP0duasBb. Zhou, L., Ye, M., Li, S., Li, N., Zhu, X., Deng, L., Liu, H., and Lei, Z. Bayesian test-time adaptation for vision- language models. InProceedings of the Computer Vision and Pattern Recognition Conference, pp. 29999–30009, 2025. 13 Adapting in the Dark: Efficient and Stable Test-Time Adaptation for Black-Bo...

  7. [7]

    Raw Logits ( z):The pre-activation output vector z∈R C, where values are unbounded ( −∞< z i <∞ ) and unnormalized

  8. [8]

    Softmax Probability Vector (p):The normalized output distribution obtained via the softmax function σ(·), such that p=σ(z)∈[0,1] C withP i pi = 1

  9. [9]

    dark knowledge

    Top-1 Hard Prediction ( ˆy):A single scalar value representing the class index with the highest confidence, ˆy= arg maxi pi, often accompanied by a single confidence score. Real-World API Protocols.To determine the most realistic setting for black-box adaptation, we analyze standard commercial Machine Learning APIs (e.g., OpenAI (Hurst et al., 2024), Clar...