pith. machine review for the scientific record. sign in

arxiv: 2605.09356 · v1 · submitted 2026-05-10 · 💻 cs.LG · cs.NI

Recognition: 2 theorem links

· Lean Theorem

Function-Space ADMM for Decentralized Federated Learning: A Control Theoretic Perspective

Authors on Pith no claims yet

Pith reviewed 2026-05-12 04:03 UTC · model grok-4.3

classification 💻 cs.LG cs.NI
keywords decentralized federated learningfunction space ADMMnon-IID dataknowledge distillationstabilization coefficientcontrol theoryconvergence analysis
0
0 comments X

The pith

Optimizing loss in function space enables faster and more stable decentralized federated learning than parameter-space methods.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces FedF-ADMM to address performance issues in decentralized federated learning caused by non-IID data distributions. It shifts the optimization to function space where loss functionals are convex, allowing the use of ADMM to derive update directions. These directions are then projected onto the parameter space through knowledge distillation. A stabilization coefficient, interpreted as a proportional-integral controller, is added for robustness in challenging non-IID cases. Experiments confirm improved convergence speed, accuracy, and consensus among devices.

Core claim

FedF-ADMM exploits the convexity of loss functionals in function space to derive ADMM-based update directions, projects them onto the parameter space via knowledge distillation, and incorporates a stabilization coefficient analyzed as a PI term to achieve faster and more stable convergence in non-IID decentralized federated learning.

What carries the argument

The ADMM updates in function space projected to parameters via knowledge distillation, with a stabilization coefficient providing PI-like control for robustness.

If this is right

  • Decentralized systems without central servers can train models effectively even when each device holds data from only a single label.
  • The control-theoretic view of the stabilization coefficient suggests ways to tune it for different network conditions.
  • Knowledge distillation acts as a reliable interface between convex function-space optimization and practical neural network parameters.
  • Overall consensus among devices improves alongside individual accuracy.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Extending this function-space approach could apply to other distributed optimization problems involving non-convex losses by using similar projections.
  • Testing the method on larger networks or different model architectures would reveal its scalability limits.
  • Integrating adaptive control strategies based on this PI interpretation might further enhance performance in dynamic environments.

Load-bearing premise

The loss functionals are convex in function space and the knowledge distillation projection does not degrade the ADMM update directions significantly.

What would settle it

Demonstrating that the projected updates fail to converge or that the stabilization coefficient causes instability in a standard non-IID FL benchmark would disprove the method's effectiveness.

Figures

Figures reproduced from arXiv: 2605.09356 by Akihito Taya, Kaoru Sezaki, Yuuki Nishiyama.

Figure 1
Figure 1. Figure 1: Geometric illustration of FedF-ADMM. FedF-ADMM determines vir [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: System model of FedF-ADMM. Devices collaboratively train their [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Updates of the output values in non-IID settings. The element-wise [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Convergence performance on Fashion-MNIST. Each device has 500 [PITH_FULL_IMAGE:figures/full_fig_p009_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Convergence performance on Fashion-MNIST. Each device has 1000 [PITH_FULL_IMAGE:figures/full_fig_p009_5.png] view at source ↗
Figure 8
Figure 8. Figure 8: Tuning results on CIFAR-10 with the 1-class distribution. FedF-ADMM outperforms the other algorithms in terms of accuracy. In addition, the [PITH_FULL_IMAGE:figures/full_fig_p010_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Convergence performance on Fashion-MNIST with Dirichlet distribu [PITH_FULL_IMAGE:figures/full_fig_p011_9.png] view at source ↗
Figure 11
Figure 11. Figure 11: Dirichlet distribution and comparison of KL divergence from the [PITH_FULL_IMAGE:figures/full_fig_p011_11.png] view at source ↗
Figure 13
Figure 13. Figure 13: Comparison of convergence performance with different shared [PITH_FULL_IMAGE:figures/full_fig_p012_13.png] view at source ↗
Figure 12
Figure 12. Figure 12: Impact of stabilization coefficient 𝜈. Thick lines represent average accuracy across devices, while thin transparent lines represent individual de￾vice accuracies. Using a non-zero 𝜈 stabilizes convergence against variations in the learning rate 𝜂 and KD coefficient 𝜌. topologies: ring, star, and random networks. The top-left panel reproduces the result in [PITH_FULL_IMAGE:figures/full_fig_p012_12.png] view at source ↗
Figure 14
Figure 14. Figure 14: Comparison of convergence performance across different network [PITH_FULL_IMAGE:figures/full_fig_p012_14.png] view at source ↗
Figure 15
Figure 15. Figure 15: Random network topologies for 𝑁 = 10 and 𝑁 = 20 devices. VII. CONCLUSION We proposed FedF-ADMM, a novel DFL algorithm based on D-ADMM in function space. The key idea of FedF￾ADMM is to optimize prediction models in function space using ADMM so as to alleviate issues arising from the non￾convexity of parameter-space optimization. KD is applied to gradually update the local parameters to align the predictio… view at source ↗
read the original abstract

Decentralized federated learning (FL) is a promising approach for training machine learning models on sensor networks, Internet of Things (IoT) devices, and other edge systems where no central server exists. While federated learning offers advantages such as preserving data privacy, it often suffers from non-independent and identically distributed (IID) data distributions across devices, which cause significant performance degradation. This issue is particularly severe when directly optimizing model parameters, because neural network training is inherently non-convex and standard convergence guarantees for convex optimization do not apply. Unlike existing decentralized FL methods that primarily operate in parameter space, we propose federated function-space alternating direction method of multipliers (FedF-ADMM). FedF-ADMM exploits the convexity of loss functionals within function space to derive alternating direction method of multipliers (ADMM)-based update directions, which are subsequently projected onto the parameter space via knowledge distillation. We further introduce a stabilization coefficient to enhance robustness under severe non-IID settings and analyze its behavior from a control-theoretic perspective by interpreting it as a proportional-integral (PI) term. Experiments under challenging non-IID scenarios, including settings where each device has data from only a single label, demonstrate that FedF-ADMM achieves faster and more stable convergence than existing decentralized FL methods, while attaining higher accuracy and better consensus among devices.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript proposes FedF-ADMM, a decentralized federated learning method that exploits convexity of loss functionals in function space to derive ADMM-based update directions. These directions are projected onto neural network parameter space via knowledge distillation. A stabilization coefficient is introduced and analyzed from a control-theoretic perspective as a proportional-integral (PI) controller. Experiments on challenging non-IID scenarios, including single-label-per-device partitions, claim faster and more stable convergence, higher accuracy, and improved consensus relative to existing decentralized FL methods.

Significance. If the central claims hold, the work offers a novel function-space perspective on decentralized FL that could improve robustness to extreme data heterogeneity on edge devices. The control-theoretic framing of the stabilization term provides a useful bridge between optimization and feedback control, with potential to inspire hybrid methods. The focus on single-label non-IID settings adds practical relevance for real-world IoT deployments.

major comments (2)
  1. [Abstract] Abstract: The claim that function-space ADMM updates can be projected onto parameter space via knowledge distillation without losing effectiveness is load-bearing for all convergence and consensus assertions. No analysis is supplied showing that the non-convex KD regression approximates the prescribed function-space direction, particularly when devices hold disjoint single-label data and local targets differ sharply; misalignment here directly threatens the consensus property ADMM is intended to enforce.
  2. [Method section (control-theoretic analysis)] Method section (control-theoretic analysis): The PI-controller interpretation of the stabilization coefficient is presented as post-hoc analysis rather than a core derivation. No equations demonstrate that the realized parameter-space updates continue to obey the linear structure assumed in the function-space derivation after the KD projection step, rendering the control-theoretic claims unsupported.
minor comments (2)
  1. [Abstract and experiments section] Abstract and experiments section: No error bars, statistical significance tests, or explicit list of baselines and hyperparameter settings are mentioned, which weakens the ability to assess the reported gains in convergence speed and accuracy.
  2. [Notation] Notation: The definition and update rule for the stabilization coefficient would benefit from an explicit equation reference to clarify its role in the ADMM iterations.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments on our work. We address each of the major comments in detail below, proposing revisions to address the concerns raised regarding the theoretical support for our claims.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The claim that function-space ADMM updates can be projected onto parameter space via knowledge distillation without losing effectiveness is load-bearing for all convergence and consensus assertions. No analysis is supplied showing that the non-convex KD regression approximates the prescribed function-space direction, particularly when devices hold disjoint single-label data and local targets differ sharply; misalignment here directly threatens the consensus property ADMM is intended to enforce.

    Authors: We agree that a formal analysis of the approximation quality of the knowledge distillation projection is absent from the manuscript, and this represents a limitation in rigorously justifying the consensus properties. The paper relies on empirical validation through experiments on single-label non-IID partitions, where FedF-ADMM shows superior performance. In revision, we will add a new subsection discussing the potential for misalignment in the projection step and its implications for ADMM's consensus guarantee. We will also include quantitative measurements of the function-space error in the experiments section to provide better support for the claims. The abstract will be revised to emphasize the empirical nature of the results. revision: yes

  2. Referee: [Method section (control-theoretic analysis)] Method section (control-theoretic analysis): The PI-controller interpretation of the stabilization coefficient is presented as post-hoc analysis rather than a core derivation. No equations demonstrate that the realized parameter-space updates continue to obey the linear structure assumed in the function-space derivation after the KD projection step, rendering the control-theoretic claims unsupported.

    Authors: The control-theoretic analysis is presented as an interpretive framework to motivate the choice of the stabilization coefficient, drawing an analogy to PI controllers for stability. We acknowledge that no explicit equations are provided showing preservation of the linear update structure post-KD projection. This is because the projection is nonlinear. We will revise the relevant section to make this post-hoc nature explicit and to clarify the assumptions under which the analogy holds approximately. We will also add a remark on the limitations of the control-theoretic view in the parameter space. revision: partial

Circularity Check

0 steps flagged

No significant circularity; derivation uses external convexity assumption and separate projection step

full rationale

The paper derives ADMM updates from the convexity of loss functionals in function space (an external mathematical property), then applies knowledge distillation as a projection to parameter space. The stabilization coefficient is introduced for robustness and given a post-hoc PI-controller interpretation from control theory, but this analysis does not feed back into the derivation or force any result by construction. No equations reduce the claimed convergence or consensus properties to fitted inputs, self-citations, or renamed empirical patterns. Experiments provide independent validation under non-IID conditions rather than tautological confirmation. The chain remains self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central claim rests on the domain assumption of convexity in function space and the effectiveness of the knowledge distillation projection step; the stabilization coefficient functions as a tunable parameter whose specific value is not derived from first principles in the abstract.

free parameters (1)
  • stabilization coefficient
    Introduced to enhance robustness under severe non-IID settings; its value is chosen or tuned based on the control-theoretic view rather than derived parameter-free.
axioms (1)
  • domain assumption Loss functionals are convex within function space
    Invoked to derive ADMM-based update directions that are then projected to parameter space.

pith-pipeline@v0.9.0 · 5544 in / 1399 out tokens · 38608 ms · 2026-05-12T04:03:50.391336+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

51 extracted references · 51 canonical work pages · 1 internal anchor

  1. [1]

    Communication-efficient learning of deep networks from decentralized data,

    B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” inProc. 20th Int. Conf. Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, Apr. 2017, pp. 1273–1282

  2. [2]

    Decentralized federated learning: Fundamentals, state-of-the-art, frameworks, trends, and challenges,

    E. T. M. Beltrán, M. Q. Pérez, P. M. S. Sánchez, S. L. Bernal, G. Bovet, M. G. Pérez, G. M. Pérez, and A. H. Celdrán, “Decentralized federated learning: Fundamentals, state-of-the-art, frameworks, trends, and challenges,”IEEE Commun. Surveys Tuts., vol. 25, no. 4, pp. 2983– 3013, 2023

  3. [3]

    IPLS: A framework for decentralized federated learning,

    C. Pappas, D. Chatzopoulos, S. Lalis, and M. Vavalis, “IPLS: A framework for decentralized federated learning,” inProc. 2021 IFIP Networking Conference (IFIP Networking), Jun. 2021, pp. 1–6

  4. [4]

    Decentralized federated learning of deep neural networks on non-IID data,

    N. Onoszko, G. Karlsson, O. Mogren, and E. L. Zec, “Decentralized federated learning of deep neural networks on non-IID data,”arXiv preprint arXiv:2107.08517 [cs.LG], 2021

  5. [5]

    Enhancing decentralized federated learning for non-IID data on heterogeneous devices,

    M. Chen, Y . Xu, H. Xu, and L. Huang, “Enhancing decentralized federated learning for non-IID data on heterogeneous devices,” inProc. 2023 IEEE 39th International Conference on Data Engineering (ICDE). Anaheim, CA, USA: IEEE, Apr. 2023, pp. 2289–2302

  6. [6]

    Decentralized and model-free federated learning: Consensus-based distillation in function space,

    A. Taya, T. Nishio, M. Morikura, and K. Yamamoto, “Decentralized and model-free federated learning: Consensus-based distillation in function space,”IEEE Trans. Signal Inf. Process. Netw., vol. 8, pp. 799–814, Sep. 2022

  7. [7]

    Can decentralized algorithms outperform centralized algorithms? a case study for decentralized parallel stochastic gradient descent,

    X. Lian, C. Zhang, H. Zhang, C.-J. Hsieh, W. Zhang, and J. Liu, “Can decentralized algorithms outperform centralized algorithms? a case study for decentralized parallel stochastic gradient descent,” inProc. 31st Conf. Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, Dec. 2017, pp. 5330–5340

  8. [8]

    Peer-to-peer federated learning on graphs,

    A. Lalitha, O. C. Kilinc, T. Javidi, and F. Koushanfar, “Peer-to-peer federated learning on graphs,”arXiv preprint arXiv:1901.11173, Jan. 2019

  9. [9]

    Federated learning with cooper- ating devices: A consensus approach for massive IoT networks,

    S. Savazzi, M. Nicoli, and V . Rampa, “Federated learning with cooper- ating devices: A consensus approach for massive IoT networks,”IEEE Internet Things J., vol. 7, no. 5, pp. 4641–4654, May 2020

  10. [10]

    Network-density-controlled de- centralized parallel stochastic gradient descent in wireless systems,

    K. Sato, Y . Satoh, and D. Sugimura, “Network-density-controlled de- centralized parallel stochastic gradient descent in wireless systems,” in Proc. IEEE Int. Conf. Commun. (ICC), Virtual Conference, Jun. 2020

  11. [11]

    Decentralized federated averaging,

    T. Sun, D. Li, and B. Wang, “Decentralized federated averaging,”IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 4, pp. 4289–4301, 2023

  12. [12]

    Collaborative deep learning in fixed topology networks,

    Z. Jiang, A. Balu, C. Hegde, and S. Sarkar, “Collaborative deep learning in fixed topology networks,” inProc. 31st Conf. Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, Dec. 2017, pp. 1–11

  13. [13]

    D- ADMM: A communication-efficient distributed algorithm for separable optimization,

    J. F. C. Mota, J. M. F. Xavier, P. M. Q. Aguiar, and M. Püschel, “D- ADMM: A communication-efficient distributed algorithm for separable optimization,”IEEE Trans. Signal Process., vol. 61, no. 10, pp. 2718– 2723, 2013. JOURNAL OF SIG.., VOL. XX, NO. XX, JANUARY 2020 15

  14. [14]

    Communication-efficient ADMM-based distributed algorithms for sparse training,

    G. Wang, Y . Lei, Y . Qiu, L. Lou, and Y . Li, “Communication-efficient ADMM-based distributed algorithms for sparse training,”Neurocomput- ing, vol. 550, p. 126456, 2023

  15. [15]

    Edge-consensus learning: Deep learning on P2P networks with nonhomogeneous data,

    K. Niwa, N. Harada, G. Zhang, and W. B. Kleijn, “Edge-consensus learning: Deep learning on P2P networks with nonhomogeneous data,” inProc. 26th ACM SIGKDD Int. Conf. Knowledge Discovery & Data Mining, Virtual Conference, Aug. 2020, pp. 668–678

  16. [16]

    DFedADMM: Dual constraint controlled model inconsistency for decentralize federated learning,

    Q. Li, L. Shen, G. Li, Q. Yin, and D. Tao, “DFedADMM: Dual constraint controlled model inconsistency for decentralize federated learning,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 47, no. 6, pp. 4803–4815, 2025

  17. [17]

    Federated learning via inexact ADMM,

    S. Zhou and G. Y . Li, “Federated learning via inexact ADMM,”IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 8, pp. 9699–9708, 2023

  18. [18]

    An inexact accelerated stochastic ADMM for separable convex optimization,

    J. Bai, W. W. Hager, and H. Zhang, “An inexact accelerated stochastic ADMM for separable convex optimization,”Computational Optimiza- tion and Applications, vol. 81, no. 2, pp. 479–518, 2022

  19. [19]

    An inexact ADMM for separable nonconvex and nonsmooth optimization,

    J. Bai, M. Zhang, and H. Zhang, “An inexact ADMM for separable nonconvex and nonsmooth optimization,”Computational Optimization and Applications, pp. 1–35, 2025

  20. [20]

    Linearized ADMM for nonconvex non- smooth optimization with convergence analysis,

    Q. Liu, X. Shen, and Y . Gu, “Linearized ADMM for nonconvex non- smooth optimization with convergence analysis,”IEEE Access, vol. 7, pp. 76 131–76 144, 2019

  21. [21]

    Knowledge distillation in federated learning: a practical guide,

    A. Mora, I. Tenison, P. Bellavista, and I. Rish, “Knowledge distillation in federated learning: a practical guide,” inProc. 33rd International Joint Conference on Artificial Intelligence, Jeju, Korea, Aug. 2024, pp. 8188–8196

  22. [22]

    arXiv preprint arXiv:1804.03235 , year=

    R. Anil, G. Pereyra, A. Passos, R. Ormandi, G. E. Dahl, and G. E. Hinton, “Large scale distributed neural network training through online distillation,”arXiv preprint arXiv:1804.03235, Apr. 2018

  23. [23]

    Deep mutual learning,

    Y . Zhang, T. Xiang, T. M. Hospedales, and H. Lu, “Deep mutual learning,” inProc. 2018 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR). IEEE, 2018, pp. 4320–4328

  24. [24]

    Distilling the Knowledge in a Neural Network

    G. Hinton, O. Vinyals, and J. Dean, “Distilling the knowledge in a neural network,”arXiv:1503.02531 [stat.ML], 2015

  25. [25]

    Communication-efficient on-device machine learning: Federated distil- lation and augmentation under non-iid private data,

    E. Jeong, S. Oh, H. Kim, J. Park, M. Bennis, and S.-L. Kim, “Communication-efficient on-device machine learning: Federated dis- tillation and augmentation under non-IID private data,”arXiv preprint arXiv:1811.11479, Nov. 2018

  26. [26]

    Multi-hop federated private data augmentation with sample compression,

    E. Jeong, S. Oh, J. Park, H. Kim, M. Bennis, and S.-L. Kim, “Multi-hop federated private data augmentation with sample compression,”arXiv preprint arXiv:1907.06426, Jul. 2019

  27. [27]

    Distillation-based semi-supervised federated learning for communication-efficient collaborative training with non-IID private data,

    S. Itahara, T. Nishio, Y . Koda, M. Morikura, and K. Ya- mamoto, “Distillation-based semi-supervised federated learning for communication-efficient collaborative training with non-IID private data,”IEEE Trans. Mobile Comput., vol. 22, no. 1, pp. 191–205, Mar. 2021

  28. [28]

    Cronus: Ro- bust and heterogeneous collaborative learning with black-box knowledge transfer,

    H. Chang, V . Shejwalkar, R. Shokri, and A. Houmansadr, “Cronus: Ro- bust and heterogeneous collaborative learning with black-box knowledge transfer,”arXiv preprint arXiv:1912.11279, Dec. 2019

  29. [29]

    Fedmd: Heterogenous federated learning via model distillation,

    D. Li and J. Wang, “FedMD: Heterogenous federated learning via model distillation,”arXiv preprint arXiv:1910.03581, Oct. 2019

  30. [30]

    Ensemble distillation for robust model fusion in federated learning,

    T. Lin, L. Kong, S. U. Stich, and M. Jaggi, “Ensemble distillation for robust model fusion in federated learning,” inProc. 33rd Conf. Neural Information Processing Systems (NeurIPS), vol. 33, Virtual Conference, Dec. 2020, pp. 2351–2363

  31. [31]

    Communication-efficient and privacy-preserving federated learning via joint knowledge distillation and differential privacy in bandwidth- constrained networks,

    G. Gad, E. Gad, Z. M. Fadlullah, M. M. Fouda, and N. Kato, “Communication-efficient and privacy-preserving federated learning via joint knowledge distillation and differential privacy in bandwidth- constrained networks,”IEEE Trans. Veh. Technol., vol. 73, no. 11, pp. 17 586–17 601, 2024

  32. [32]

    FKD-Med: privacy-aware, communication-optimized medical image segmentation via federated learning and model lightweighting through knowledge distillation,

    G. Sun, H. Shu, F. Shao, T. Racharak, W. Kong, Y . Pan, J. Dong, S. Wang, L.-M. Nguyen, and J. Xin, “FKD-Med: privacy-aware, communication-optimized medical image segmentation via federated learning and model lightweighting through knowledge distillation,”IEEE Access, vol. 12, pp. 33 687–33 704, 2024

  33. [33]

    PFedKD: Personalized federated learning via knowledge distillation using unlabeled pseudo data for Internet of Things,

    H. Li, G. Chen, B. Wang, Z. Chen, Y . Zhu, F. Hu, J. Dai, and W. Wang, “PFedKD: Personalized federated learning via knowledge distillation using unlabeled pseudo data for Internet of Things,”IEEE Internet of Things Journal, vol. 12, no. 11, pp. 16 314–16 324, 2025

  34. [34]

    DFLStar: A decentralized federated learning framework with self-knowledge distil- lation and participant selection,

    B. Soltani, V . Haghighi, Y . Zhou, Q. Z. Sheng, and L. Yao, “DFLStar: A decentralized federated learning framework with self-knowledge distil- lation and participant selection,” inProc. of the 33rd ACM International Conference on Information and Knowledge Management, ser. CIKM ’24, Boise, ID, USA, Oct. 2024, pp. 2108–2117

  35. [35]

    DFUN- KDF: Efficient and robust decentralized federated framework for UA V networks via knowledge distillation and filtering,

    W. Yang, Y . Liu, X. Leng, H. Gu, G. Jiang, X. Yu, and X. Cao, “DFUN- KDF: Efficient and robust decentralized federated framework for UA V networks via knowledge distillation and filtering,”Communications in Transportation Research, vol. 5, p. 100173, 2025

  36. [36]

    Federated learning with differential pri- vacy: Algorithms and performance analysis,

    K. Wei, J. Li, M. Ding, C. Ma, H. H. Yang, F. Farokhi, S. Jin, T. Q. S. Quek, and H. Vincent Poor, “Federated learning with differential pri- vacy: Algorithms and performance analysis,”IEEE Trans. Inf. Forensics Security, vol. 15, pp. 3454–3469, 2020

  37. [37]

    Clustered federated learning with adaptive local differential privacy on heterogeneous IoT data,

    Z. He, L. Wang, and Z. Cai, “Clustered federated learning with adaptive local differential privacy on heterogeneous IoT data,”IEEE Internet Things J., vol. 11, no. 1, pp. 137–146, 2024

  38. [38]

    A framework for privacy-preserving in IoV using federated learning with differential privacy,

    M. Adnan, M. H. Syed, A. Anjum, and S. Rehman, “A framework for privacy-preserving in IoV using federated learning with differential privacy,”IEEE Access, vol. 13, pp. 13 507–13 521, 2025

  39. [39]

    Data-free knowledge distillation for het- erogeneous federated learning,

    Z. Zhu, J. Hong, and J. Zhou, “Data-free knowledge distillation for het- erogeneous federated learning,” inProc. 38th International Conference on Machine Learning, vol. 139, Jul. 2021, pp. 12 878–12 889

  40. [40]

    FedGMKD: An efficient prototype feder- ated learning framework through knowledge distillation and discrepancy- aware aggregation,

    J. Zhang, C. Shan, and J. Han, “FedGMKD: An efficient prototype feder- ated learning framework through knowledge distillation and discrepancy- aware aggregation,” inAdvances in Neural Information Processing Systems, vol. 37, 2024, pp. 118 326–118 356

  41. [41]

    Fine-tuning global model via data-free knowledge distillation for non-IID federated learning,

    L. Zhang, L. Shen, L. Ding, D. Tao, and L.-Y . Duan, “Fine-tuning global model via data-free knowledge distillation for non-IID federated learning,” inProc. 2022 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 2022, pp. 10 164–10 173

  42. [42]

    Extending control theory into federated learning data heterogeneity problem,

    O. Mashaal, S. Baadel, and H. AbouZied, “Extending control theory into federated learning data heterogeneity problem,” in2024 IEEE International Conference on Artificial Intelligence and Mechatronics Systems (AIMS). Bandung, Indonesia: IEEE, Feb. 2024, pp. 1–4

  43. [43]

    SCAFFOLD: Stochastic controlled averaging for federated learning,

    S. P. Karimireddy, S. Kale, M. Mohri, S. Reddi, S. Stich, and A. T. Suresh, “SCAFFOLD: Stochastic controlled averaging for federated learning,” inProc. 37th International Conference on Machine Learning. Virtual Conference: PMLR, July 2020, pp. 5132–5143

  44. [44]

    Fed-PID: An adaptive learning rate scheduler for federated learning with PID controllers,

    S. Li, M. Dai, S. Kianoush, A. Minora, and S. Savazzi, “Fed-PID: An adaptive learning rate scheduler for federated learning with PID controllers,”IEEE Trans. Cogn. Commun. Netw., pp. 1–13, 2025

  45. [45]

    FedADT: An adaptive method based on derivative term for federated learning,

    H. Gao, Q. Wu, X. Zhao, J. Zhu, and M. Zhang, “FedADT: An adaptive method based on derivative term for federated learning,”Sensors, vol. 23, no. 13, p. 6034, 2023

  46. [46]

    FedPIDAvg: A PID controller inspired aggregation method for federated learning,

    L. Mächler, I. Ezhov, S. Shit, and J. C. Paetzold, “FedPIDAvg: A PID controller inspired aggregation method for federated learning,” in Proc. Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 8th International Workshop, BrainLes 2022, Held in Conjunction with MICCAI 2022. Singapore: Springer, Sep. 2022, pp. 209–217

  47. [47]

    Federated learning for collaborative controller design of connected and autonomous vehicles,

    T. Zeng, O. Semiari, M. Chen, W. Saad, and M. Bennis, “Federated learning for collaborative controller design of connected and autonomous vehicles,” in60th IEEE Conference on Decision and Control (CDC). Austin, TX, USA: IEEE, Dec. 2021, pp. 5033–5038

  48. [48]

    Advances and open problems in federated learning,

    P. Kairouzet al., “Advances and open problems in federated learning,” Foundations and Trends® in Machine Learning, vol. 14, no. 1–2, pp. 1–210, Jun. 2021

  49. [49]

    Federated optimization in heterogeneous networks,

    T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V . Smith, “Federated optimization in heterogeneous networks,” inProc. Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V . Sze, Eds., vol. 2, 2020, pp. 429–450

  50. [50]

    Distributed optimization and statistical learning via the alternating direction method of multipliers,

    S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Ecksteinet al., “Distributed optimization and statistical learning via the alternating direction method of multipliers,”Foundations and Trends® in Machine learning, vol. 3, no. 1, pp. 1–122, 2011

  51. [51]

    Optuna: A next-generation hyperparameter optimization framework,

    T. Akiba, S. Sano, T. Yanase, T. Ohta, and M. Koyama, “Optuna: A next-generation hyperparameter optimization framework,” inProc. the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ser. KDD ’19, Anchorage, AK, USA, Jul. 2019, pp. 2623–2631. JOURNAL OF SIG.., VOL. XX, NO. XX, JANUARY 2020 16 Akihito Tayareceived the B.E. degr...