{"total":10,"items":[{"citing_arxiv_id":"2605.13434","ref_index":54,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"Rescaled Asynchronous SGD: Optimal Distributed Optimization under Data and System Heterogeneity","primary_cat":"cs.LG","submitted_at":"2026-05-13T12:27:22+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"Rescaled ASGD recovers convergence to the true global objective by rescaling worker stepsizes proportional to computation times, matching the known time lower bound in the leading term under non-convex smoothness and bounded heterogeneity.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.11915","ref_index":41,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"ISAC for AI: A Trade-off Framework Across Data Acquisition and Transfer in Federated Learning","primary_cat":"eess.SP","submitted_at":"2026-05-12T10:30:13+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":4.0,"formal_verification":"none","one_line_summary":"A closed-form FL convergence upper bound incorporating sensing SNR, dataset size, and transmission reliability enables joint optimization of sensing power, snapshots, and communication power in ISAC systems.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.11857","ref_index":25,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"Beyond Parameter Aggregation: Semantic Consensus for Federated Fine-Tuning of LLMs","primary_cat":"cs.LG","submitted_at":"2026-05-12T09:41:40+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":7.0,"formal_verification":"none","one_line_summary":"Semantic consensus on model outputs for public prompts enables federated LLM fine-tuning that matches parameter-aggregation baselines with orders-of-magnitude lower communication.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.09144","ref_index":39,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"FedVSSAM: Mitigating Flatness Incompatibility in Sharpness-Aware Federated Learning","primary_cat":"cs.LG","submitted_at":"2026-05-09T20:03:02+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"FedVSSAM mitigates flatness incompatibility in SAM-based federated learning by consistently using a variance-suppressed adjusted direction for local perturbation, descent, and global updates, with non-convex convergence guarantees.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.08760","ref_index":10,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"FedGMI: Generative Model-Driven Federated Learning for Probabilistic Mixture Inference","primary_cat":"cs.LG","submitted_at":"2026-05-09T07:45:10+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":5.0,"formal_verification":"none","one_line_summary":"FedGMI applies VAEs as density estimators in federated learning to infer mixture proportions of shared distributions for structured personalization under data heterogeneity.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.02337","ref_index":22,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"FedPLT: Scalable, Resource-Efficient, and Heterogeneity-Aware Federated Learning via Partial Layer Training","primary_cat":"cs.DC","submitted_at":"2026-05-04T08:41:02+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":5.0,"formal_verification":"none","one_line_summary":"FedPLT assigns client-specific model layers for training and matches or beats full-model federated learning accuracy with 71-82 percent fewer trainable parameters per client.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2604.27434","ref_index":22,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"AdaBFL: Multi-Layer Defensive Adaptive Aggregation for Bzantine-Robust Federated Learning","primary_cat":"cs.LG","submitted_at":"2026-04-30T05:18:44+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":5.0,"formal_verification":"none","one_line_summary":"AdaBFL introduces a novel three-layer adaptive aggregation mechanism for Byzantine-robust federated learning that counters complex attacks, provides non-convex non-iid convergence guarantees, and shows superior performance in experiments.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2604.26388","ref_index":29,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"SplitFT: An Adaptive Federated Split Learning System For LLMs Fine-Tuning","primary_cat":"cs.DC","submitted_at":"2026-04-29T07:58:10+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":5.0,"formal_verification":"none","one_line_summary":"SplitFT adapts cut-layer selection and reduces LoRA rank per client in federated split learning to improve efficiency and performance when fine-tuning LLMs on heterogeneous devices and data.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2604.19118","ref_index":114,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"DP-FlogTinyLLM: Differentially private federated log anomaly detection using Tiny LLMs","primary_cat":"cs.CR","submitted_at":"2026-04-21T05:56:51+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":4.0,"formal_verification":"none","one_line_summary":"DP-FLogTinyLLM combines federated learning, differential privacy, and LoRA-tuned tiny LLMs to match centralized log anomaly detection performance on Thunderbird and BGL datasets while preserving privacy.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2604.06819","ref_index":20,"ref_count":1,"confidence":0.9,"is_internal_anchor":false,"paper_title":"Beyond End-to-End: Dynamic Chain Optimization for Private LLM Adaptation on the Edge","primary_cat":"cs.DC","submitted_at":"2026-04-08T08:37:17+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"ChainFed achieves memory-efficient private LLM fine-tuning on edge devices through sequential layer-by-layer adapter training with dynamic co-tuning, perceptive optimization, and adaptive starting point selection, improving accuracy by up to 46.46%.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null}],"limit":50,"offset":0}