pith. machine review for the scientific record. sign in

arxiv: 1811.03604 · v2 · submitted 2018-11-08 · 💻 cs.CL

Recognition: unknown

Federated Learning for Mobile Keyboard Prediction

Andrew Hard, Chlo\'e Kiddon, Daniel Ramage, Fran\c{c}oise Beaufays, Hubert Eichner, Kanishka Rao, Rajiv Mathews, Sean Augenstein, Swaroop Ramaswamy

Authors on Pith no claims yet
classification 💻 cs.CL
keywords federatedtraininglearningclientdevicespredictionalgorithmdata
0
0 comments X
read the original abstract

We train a recurrent neural network language model using a distributed, on-device learning framework called federated learning for the purpose of next-word prediction in a virtual keyboard for smartphones. Server-based training using stochastic gradient descent is compared with training on client devices using the Federated Averaging algorithm. The federated algorithm, which enables training on a higher-quality dataset for this use case, is shown to achieve better prediction recall. This work demonstrates the feasibility and benefit of training language models on client devices without exporting sensitive user data to servers. The federated learning environment gives users greater control over the use of their data and simplifies the task of incorporating privacy by default with distributed training and aggregation across a population of client devices.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 9 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. When More Parameters Hurt: Foundation Model Priors Amplify Worst-Client Disparity Under Extreme Federated Heterogeneity

    cs.LG 2026-05 unverdicted novelty 7.0

    Foundation model priors amplify worst-client disparity under extreme federated heterogeneity, creating a fairness paradox where larger models perform worse for disadvantaged clients.

  2. Unified Compression Algorithm for Distributed Nonconvex Optimization: Generalized to 1-Bit, Saturation, and Bounded Noise

    math.OC 2026-04 unverdicted novelty 7.0

    A unified compression algorithm for distributed nonconvex optimization achieves O(1/sqrt(T)) convergence for locally-bounded compressors, matching centralized 1-bit methods, with an improved O(1/T^{2/3}) rate after on...

  3. XFED: Non-Collusive Model Poisoning Attack Against Byzantine-Robust Federated Classifiers

    cs.CR 2026-04 unverdicted novelty 7.0

    XFED is the first aggregation-agnostic non-collusive model poisoning attack that bypasses eight state-of-the-art defenses on six benchmark datasets without attacker coordination.

  4. Distributed Online Convex Optimization with Compressed Communication: Optimal Regret and Applications

    cs.LG 2026-04 unverdicted novelty 7.0

    Optimal regret bounds O(δ^{-1/2}√T) for convex and O(δ^{-1} log T) for strongly convex losses are achieved in distributed online convex optimization under compressed communication.

  5. Beyond Corner Patches: Semantics-Aware Backdoor Attack in Federated Learning

    cs.CR 2026-03 unverdicted novelty 7.0

    SABLE shows that semantics-aware natural triggers enable effective backdoor attacks in federated learning against multiple aggregation rules while preserving benign accuracy.

  6. Rescaled Asynchronous SGD: Optimal Distributed Optimization under Data and System Heterogeneity

    cs.LG 2026-05 unverdicted novelty 6.0

    Rescaled ASGD recovers convergence to the true global objective by rescaling worker stepsizes proportional to computation times, matching the known time lower bound in the leading term under non-convex smoothness and ...

  7. Analytically Characterized Optimal Power Control for Signal-Level-Integrated Sensing, Computing and Communication in Federated Learning

    cs.IT 2026-04 unverdicted novelty 6.0

    An optimal convex-reformulated power control algorithm is derived for signal-level integrated sensing, computing and communication in AirComp-based federated learning under a joint target detection constraint.

  8. HUOZIIME: An On-Device LLM-enhanced Input Method for Deep Personalization

    cs.CL 2026-03 unverdicted novelty 5.0

    HUOZIIME is an on-device LLM-powered input method with post-training on synthesized data and hierarchical memory that achieves efficient execution and memory-driven personalization.

  9. Split and Aggregation Learning for Foundation Models Over Mobile Embodied AI Network (MEAN): A Comprehensive Survey

    cs.IT 2026-05 unverdicted novelty 3.0

    The paper surveys split and aggregation learning for foundation models in 6G networks to improve efficiency, resource use, and data privacy in distributed AI.