pith. machine review for the scientific record. sign in

arxiv: 2604.27510 · v1 · submitted 2026-04-30 · 💻 cs.LG · cs.CV

Recognition: unknown

FMCL: Class-Aware Client Clustering with Foundation Model Representations for Heterogeneous Federated Learning

Laura J. Brattain, Mahad Ali

Authors on Pith no claims yet

Pith reviewed 2026-05-07 10:07 UTC · model grok-4.3

classification 💻 cs.LG cs.CV
keywords Federated LearningClient ClusteringFoundation ModelsHeterogeneous DataNon-IID PartitioningClass-Aware EmbeddingsOne-Shot Clustering
0
0 comments X

The pith

FMCL uses class-level prototypes from a frozen foundation model to perform one-shot client clustering before federated training begins.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces FMCL to address performance drops in federated learning when client data follows different distributions. It extracts per-class embedding vectors for each client's local data using an unchanged foundation model, then measures pairwise client similarity with cosine distance on those vectors. Clustering occurs once upfront, after which separate models train within each group without further coordination or architecture changes. Experiments on standard non-IID benchmarks show higher accuracy and steadier cluster assignments than earlier methods that rely on gradients, parameters, or raw statistics.

Core claim

FMCL builds semantic client signatures by computing class-level embedding prototypes with a frozen foundation model, then applies one-shot clustering via cosine similarity on these prototypes prior to federated optimization, yielding improved accuracy and more stable groupings under non-identically distributed data.

What carries the argument

Class-level embedding prototypes extracted by a frozen foundation model, with cosine distance serving as the similarity metric for one-shot client clustering.

If this is right

  • Clustering adds no communication rounds during the federated training phase itself.
  • The approach works with any downstream model architecture since it never inspects or alters model parameters.
  • Accuracy gains appear on multiple heterogeneous benchmarks relative to prior clustering techniques.
  • Cluster assignments remain more consistent than those derived from raw data statistics or model updates.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same frozen-model signatures could support client grouping in other distributed settings that lack shared labels.
  • Re-running the one-shot step with an updated foundation model would require only local recomputation and no extra rounds.
  • Hybrid pipelines might combine these fixed prototypes with light client-specific adaptation when domain gaps are large.
  • The method could apply to privacy-sensitive scenarios where semantic grouping is needed without exchanging raw examples.

Load-bearing premise

That representations from a frozen foundation model capture class-level semantic structure reliably enough across heterogeneous domains for cosine similarity to produce useful client clusters without any adaptation.

What would settle it

An experiment on a heterogeneous benchmark where the foundation-model prototypes produce clusters that either fail to raise accuracy or change substantially across repeated runs compared with existing clustering baselines.

Figures

Figures reproduced from arXiv: 2604.27510 by Laura J. Brattain, Mahad Ali.

Figure 1
Figure 1. Figure 1: The FMCL Framework. (Left) Clients extract class-aware semantic signa￾tures using a frozen domain-specific foundation model. (Center) The server computes an overlap-aware cosine distance matrix and selects the number of clusters K via CV-guided silhouette analysis. (Right) One-shot clustering is followed by independent federated training within each cluster. 2 Related Work Federated Learning under Statisti… view at source ↗
Figure 2
Figure 2. Figure 2: Performance comparison of federated learning strategies across three datasets. FMCL achieves higher accuracy and faster convergence than competing methods. We also report results using our automatic cluster selection algorithm (Algo￾rithm 5). Across all datasets, Auto-K further improves performance and reduces variance compared to fixed-K clustering, demonstrating the robustness of our cluster selection st… view at source ↗
read the original abstract

Federated Learning (FL) enables collaborative model training across distributed clients without sharing raw data, yet its performance deteriorates under statistical heterogeneity. Clustered Federated Learning addresses this challenge by grouping similar clients and training separate models per cluster. However, existing clustering strategies often rely on raw data statistics, model parameters, or heuristic similarity measures that fail to capture class-level semantic structure across heterogeneous domains and frequently require iterative coordination. We propose FMCL, a one-shot, class-aware client clustering framework that leverages foundation model representations to construct semantic client signatures. Using a frozen foundation model, FMCL computes class-level embedding prototypes for each client and measures similarity via cosine distance between their class-aware representations. Clustering is performed once prior to training, introducing no additional communication during federated optimization and remaining agnostic to the downstream model architecture. Extensive experiments across heterogeneous benchmarks demonstrate that FMCL improves federated performance and yields more stable clustering behavior compared to existing clustering-based methods under non-identically distributed data partitioning.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript introduces FMCL, a one-shot client clustering framework for heterogeneous federated learning. It extracts class-level embedding prototypes from each client's local data using a frozen foundation model, measures inter-client similarity via cosine distance on these class-aware representations, and performs clustering once before federated training begins. The method is presented as architecture-agnostic and communication-free during optimization. The central claim is that this yields improved federated accuracy and more stable clusters than prior clustering-based FL approaches on non-IID benchmarks.

Significance. If the empirical results hold after proper validation, FMCL would offer a lightweight, pre-training-leveraging alternative to parameter- or heuristic-based clustering in FL, reducing coordination overhead while addressing statistical heterogeneity through semantic client signatures derived from general-purpose models.

major comments (2)
  1. [Abstract] Abstract: The central claim of performance gains and improved stability is asserted without any quantitative results, baseline comparisons, statistical tests, ablation studies, or experimental protocol details, leaving the empirical contribution unsupported at the level needed for assessment.
  2. [Method] Method (clustering step): The one-shot cosine clustering on frozen FM class prototypes assumes that these embeddings reliably capture class semantics and produce meaningful client groups across heterogeneous partitions without domain adaptation or refinement; no analysis, sensitivity checks, or validation on out-of-distribution client data is provided to support this load-bearing assumption.
minor comments (2)
  1. [Method] Clarify how class prototypes are exactly computed (e.g., averaging, weighting) and whether the foundation model choice is fixed or explored.
  2. [Experiments] Ensure all figures and tables in the experimental section include error bars, multiple random seeds, and direct numerical comparisons to the cited baselines.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback. We agree that the abstract would benefit from more concrete empirical support and that additional analysis would strengthen the validation of the clustering assumption. We address each major comment below and will revise the manuscript accordingly.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim of performance gains and improved stability is asserted without any quantitative results, baseline comparisons, statistical tests, ablation studies, or experimental protocol details, leaving the empirical contribution unsupported at the level needed for assessment.

    Authors: We agree that the abstract presents the claims at a high level. In the revised version, we will update the abstract to include key quantitative highlights from our experiments, such as the observed improvements in federated accuracy and measures of clustering stability across non-IID benchmarks. The full experimental protocol, baseline comparisons (including methods such as IFCA and parameter-based clustering), ablation studies, and any statistical tests are already provided in detail in Sections 4 and 5. This change will better align the abstract with the empirical content of the paper. revision: yes

  2. Referee: [Method] Method (clustering step): The one-shot cosine clustering on frozen FM class prototypes assumes that these embeddings reliably capture class semantics and produce meaningful client groups across heterogeneous partitions without domain adaptation or refinement; no analysis, sensitivity checks, or validation on out-of-distribution client data is provided to support this load-bearing assumption.

    Authors: The assumption is indeed central to the approach. Our experiments on multiple heterogeneous benchmarks demonstrate that the resulting clusters yield performance gains and greater stability relative to prior methods, providing indirect empirical support for the semantic quality of the frozen foundation model prototypes. We acknowledge the absence of explicit sensitivity analysis or dedicated out-of-distribution validation. In the revision, we will add sensitivity checks on foundation model choice and further discussion of robustness under distribution shifts. The non-IID partitions used already incorporate substantial heterogeneity that serves as a practical test of the assumption. revision: partial

Circularity Check

0 steps flagged

No significant circularity in derivation chain

full rationale

The paper presents FMCL as a procedural pipeline: a frozen foundation model computes class-level embedding prototypes per client, cosine similarity is applied once to form clusters before any federated training, and the process is architecture-agnostic with no additional communication. No equations, fitted parameters, or predictions are described that reduce the claimed performance gains to the inputs by construction. The abstract and method summary contain no self-citations of prior author results used as load-bearing uniqueness theorems, no smuggled ansatzes, and no renaming of known results as novel derivations. The central claim rests on direct computation from external foundation-model representations and empirical benchmarks, making the approach self-contained against external data without circular reductions.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on the domain assumption that foundation-model embeddings encode transferable class semantics and on the procedural choice of one-shot cosine clustering; no free parameters are explicitly fitted in the abstract description, and no new entities are postulated.

axioms (1)
  • domain assumption Frozen foundation model representations preserve class-level semantic similarities across heterogeneous client domains
    Invoked to justify using cosine distance on class prototypes as a proxy for client similarity without further training or adaptation.

pith-pipeline@v0.9.0 · 5469 in / 1406 out tokens · 100678 ms · 2026-05-07T10:07:58.560047+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

16 extracted references · 8 canonical work pages

  1. [1]

    In: In- ternational Conference on Learning Representations (2021),https://openreview

    Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale. In: In- ternational Conference on Learning Representations (2021),https://openreview. net/forum?id=YicbFdNTTy

  2. [2]

    In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H

    Fallah, A., Mokhtari, A., Ozdaglar, A.: Personalized federated learning with the- oretical guarantees: A model-agnostic meta-learning approach. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (eds.) Advances in Neu- ral Information Processing Systems. vol. 33, pp. 3557–3568. Curran Associates, Inc. (2020),https://proceedings.neurips.cc/p...

  3. [3]

    IEEE Transactions on Information Theory68(12), 8076– 8091 (2022).https://doi.org/10.1109/TIT.2022.3192506

    Ghosh, A., Chung, J., Yin, D., Ramchandran, K.: An efficient framework for clus- tered federated learning. IEEE Transactions on Information Theory68(12), 8076– 8091 (2022).https://doi.org/10.1109/TIT.2022.3192506

  4. [4]

    Hsu, T.M.H., Qi, H., Brown, M.: Measuring the effects of non-identical data dis- tribution for federated visual classification (2019),https://arxiv.org/abs/1909. 06335

  5. [5]

    Pattern Recog- nition153, 110500 (2024).https://doi.org/https://doi.org/10.1016/j

    Jiao, J., Zhou, J., Li, X., Xia, M., Huang, Y., Huang, L., Wang, N., Zhang, X., Zhou, S., Wang, Y., Guo, Y.: Usfm: A universal ultrasound foundation model generalized to tasks and organs towards label efficient image analysis. Medical Im- age Analysis96, 103202 (2024).https://doi.org/https://doi.org/10.1016/j. media.2024.103202,https://www.sciencedirect.c...

  6. [6]

    In: International Conference on Machine Learning

    Karimireddy, S.P., Kale, S., Mohri, M., Reddi, S., Stich, S., Suresh, A.T.: Scaffold: Stochastic controlled averaging for federated learning. In: International Conference on Machine Learning. pp. 5132–5143. PMLR (2020)

  7. [7]

    Improving pronunciation and accent conversion through knowledge distillation and synthetic ground-truth from native tts

    Lao, G., Zhang, X., Li, Y., Gong, Y.J.: Lightweight clustered federated learning via feature extraction. In: ICASSP 2025 - 2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). pp. 1–5 (2025).https: //doi.org/10.1109/ICASSP49660.2025.10889437

  8. [8]

    Proceedings of Machine learning and sys- tems2, 429–450 (2020)

    Li, T., Sahu, A.K., Zaheer, M., Sanjabi, M., Talwalkar, A., Smith, V.: Federated optimization in heterogeneous networks. Proceedings of Machine learning and sys- tems2, 429–450 (2020)

  9. [9]

    In: Singh, A., Zhu, J

    McMahan, B., Moore, E., Ramage, D., Hampson, S., Arcas, B.A.y.: Communication-Efficient Learning of Deep Networks from Decentralized Data. In: Singh, A., Zhu, J. (eds.) Proceedings of the 20th International Conference on Artifi- cial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 54, pp. 1273–1282. PMLR (20–22 Apr 2017),https:...

  10. [10]

    Adaptive federated optimization,

    Reddi, S., Charles, Z., Zaheer, M., Garrett, Z., Rush, K., Konečn` y, J., Ku- mar, S., McMahan, H.B.: Adaptive federated optimization. arXiv preprint arXiv:2003.00295 (2020)

  11. [11]

    IEEE Transactions on Neural Networks and Learning Systems32(8), 3710–3722 (2021).https:// doi.org/10.1109/TNNLS.2020.3015958,http://dx.doi.org/10.1109/TNNLS

    Sattler, F., Müller, K.R., Samek, W.: Clustered federated learning: Model-agnostic distributed multi-task optimization under privacy constraints. IEEE Transactions on Neural Networks and Learning Systems32(8), 3710–3722 (2021).https:// doi.org/10.1109/TNNLS.2020.3015958,http://dx.doi.org/10.1109/TNNLS. 2020.3015958 16 M. Ali et al

  12. [12]

    Vahidian, S., Morafah, M., Wang, W., Kungurtsev, V., Chen, C., Shah, M., Lin, B.: Efficient distribution similarity identification in clustered federated learning via principal angles between client data subspaces. In: Proceedings of the Thirty- Seventh AAAI Conference on Artificial Intelligence and Thirty-Fifth Conference on Innovative Applications of Ar...

  13. [13]

    In: Proceedings of the 34th International Conference on Neural Information Processing Systems

    Wang, J., Liu, Q., Liang, H., Joshi, G., Poor, H.V.: Tackling the objective incon- sistency problem in heterogeneous federated optimization. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. NIPS ’20, Curran Associates Inc., Red Hook, NY, USA (2020)

  14. [14]

    Nature634(8035), 970–978 (Oct 2024)

    Wang, X., Zhao, J., Marostica, E., Yuan, W., Jin, J., Zhang, J., Li, R., Tang, H., Wang, K., Li, Y., Wang, F., Peng, Y., Zhu, J., Zhang, J., Jackson, C.R., Zhang, J., Dillon,D.,Lin,N.U.,Sholl,L.,Denize,T.,Meredith,D.,Ligon,K.L.,Signoretti,S., Ogino, S., Golden, J.A., Nasrallah, M.P., Han, X., Yang, S., Yu, K.H.: A pathology foundation model for cancer dia...

  15. [15]

    Yu, T., Bagdasaryan, E., Shmatikov, V.: Salvaging federated learning by local adaptation (2022),https://arxiv.org/abs/2002.04758

  16. [16]

    Zhao, Y., Li, M., Lai, L., Suda, N., Civin, D., Chandra, V.: Federated learning with non-iiddata.CoRRabs/1806.00582(2018),http://arxiv.org/abs/1806.00582