PE-MAMoE combines sparsely gated mixture-of-experts actors with a non-parametric phase controller in MAPPO to maintain plasticity under dynamic user mobility and traffic, yielding 26.3% higher normalized IQM return in simulations.
Title resolution pending
3 Pith papers cite this work. Polarity classification is still indexing.
years
2026 3representative citing papers
LQL turns n-step action-sequence lower bounds into a practical hinge-loss stabilizer for off-policy Q-learning without extra networks or forward passes.
A tutorial playbook that organizes statistical evaluation into a workflow of claim, hypothesis, unit of analysis, baselines, sweeps, uncertainty, validation, and reporting, illustrated with Python code and a job-scheduling example.
citing papers explorer
-
Plasticity-Enhanced Multi-Agent Mixture of Experts for Dynamic Objective Adaptation in UAVs-Assisted Emergency Communication Networks
PE-MAMoE combines sparsely gated mixture-of-experts actors with a non-parametric phase controller in MAPPO to maintain plasticity under dynamic user mobility and traffic, yielding 26.3% higher normalized IQM return in simulations.
-
Long-Horizon Q-Learning: Accurate Value Learning via n-Step Inequalities
LQL turns n-step action-sequence lower bounds into a practical hinge-loss stabilizer for off-policy Q-learning without extra networks or forward passes.
-
How to Do Statistical Evaluations in ECE/CS Papers: A Practical Playbook for Defensible Results
A tutorial playbook that organizes statistical evaluation into a workflow of claim, hypothesis, unit of analysis, baselines, sweeps, uncertainty, validation, and reporting, illustrated with Python code and a job-scheduling example.