Prefix-tuning matches or exceeds fine-tuning on NLG tasks by optimizing a continuous prefix using 0.1% of parameters while keeping the LM frozen.
hub
Communication-efficient learning of deep networks from decentralized data
12 Pith papers cite this work. Polarity classification is still indexing.
hub tools
representative citing papers
A multimodal graph learning method for V2X beam alignment cuts overhead by over 90% and outperforms prior federated learning baselines under label and modality imbalance.
Decoupled DiLoCo enables asynchronous distributed pre-training with zero global downtime under simulated failures while preserving competitive performance on text and vision tasks.
Rennala MVR improves time complexity over Rennala SGD for smooth nonconvex stochastic optimization in heterogeneous parallel systems under a mean-squared smoothness assumption.
Experiments on real industrial time series show that partial model sharing improves diffusion model performance in bandwidth-limited non-IID settings, while full sharing stabilizes GAN training but offers less robustness than VAE or DDPM alternatives.
Federated learning on 310 CT scans from two centers yields pediatric OAR segmentation models with better cross-center robustness than local models for nine evaluated structures.
SplitFT adapts cut-layer selection and reduces LoRA rank per client in federated split learning to improve efficiency and performance when fine-tuning LLMs on heterogeneous devices and data.
DDP-SA combines client-side Laplace noise perturbation with full-threshold additive secret sharing to let federated learning servers reconstruct only aggregated noisy gradients without exposing individual client updates.
Life-logging video streams create an inevitable privacy-utility trade-off that is a foundational challenge for always-on AI systems.
FedProx outperforms FedAvg for deeper models under data heterogeneity, BSP reaches near-centralized accuracy at high communication cost, and LeNet gives the best accuracy-communication trade-off on the UC Merced dataset.
Iterative joint learning-optimization framework with convergent algorithms for pseudoconvex objectives in operational decision systems.
The paper surveys split and aggregation learning for foundation models in 6G networks to improve efficiency, resource use, and data privacy in distributed AI.
citing papers explorer
-
Prefix-Tuning: Optimizing Continuous Prompts for Generation
Prefix-tuning matches or exceeds fine-tuning on NLG tasks by optimizing a continuous prefix using 0.1% of parameters while keeping the LM frozen.
-
Scalable Multimodal Beam Alignment in V2X: An Anti-Imbalance Graph Learning Approach
A multimodal graph learning method for V2X beam alignment cuts overhead by over 90% and outperforms prior federated learning baselines under label and modality imbalance.
-
Decoupled DiLoCo for Resilient Distributed Pre-training
Decoupled DiLoCo enables asynchronous distributed pre-training with zero global downtime under simulated failures while preserving competitive performance on text and vision tasks.
-
Rennala MVR: Improved Time Complexity for Parallel Stochastic Optimization via Momentum-Based Variance Reduction
Rennala MVR improves time complexity over Rennala SGD for smooth nonconvex stochastic optimization in heterogeneous parallel systems under a mean-squared smoothness assumption.
-
On the Tradeoffs of On-Device Generative Models in Federated Predictive Maintenance Systems
Experiments on real industrial time series show that partial model sharing improves diffusion model performance in bandwidth-limited non-IID settings, while full sharing stabilizes GAN training but offers less robustness than VAE or DDPM alternatives.
-
Overcoming data scarcity through multi-center federated learning for organs-at-risk segmentation in pediatric upper abdominal radiotherapy
Federated learning on 310 CT scans from two centers yields pediatric OAR segmentation models with better cross-center robustness than local models for nine evaluated structures.
-
SplitFT: An Adaptive Federated Split Learning System For LLMs Fine-Tuning
SplitFT adapts cut-layer selection and reduces LoRA rank per client in federated split learning to improve efficiency and performance when fine-tuning LLMs on heterogeneous devices and data.
-
DDP-SA: Scalable Privacy-Preserving Federated Learning via Distributed Differential Privacy and Secure Aggregation
DDP-SA combines client-side Laplace noise perturbation with full-threshold additive secret sharing to let federated learning servers reconstruct only aggregated noisy gradients without exposing individual client updates.
-
Position: Life-Logging Video Streams Make the Privacy-Utility Trade-off Inevitable
Life-logging video streams create an inevitable privacy-utility trade-off that is a foundational challenge for always-on AI systems.
-
The Impact of Federated Learning on Distributed Remote Sensing Archives
FedProx outperforms FedAvg for deeper models under data heterogeneity, BSP reaches near-centralized accuracy at high communication cost, and LeNet gives the best accuracy-communication trade-off on the UC Merced dataset.
-
Pseudoconvex Problems in Operational Decision Systems: Algorithms for Joint Learning and Optimization
Iterative joint learning-optimization framework with convergent algorithms for pseudoconvex objectives in operational decision systems.
-
Split and Aggregation Learning for Foundation Models Over Mobile Embodied AI Network (MEAN): A Comprehensive Survey
The paper surveys split and aggregation learning for foundation models in 6G networks to improve efficiency, resource use, and data privacy in distributed AI.