In this work, we study the asymptotic behavior of mixture of experts (MoE) trained via gradient flow on supervised learning problems. Our main result establishes the propagation of chaos for a MoE as the number of experts diverges. We demonstrate that the corresponding empirical measure of their parameters is close to a probability measure that solves a nonlinear continuity equation, and we provide an explicit convergence rate that depends solely on the number of experts and on the dimensionality of the parameter space. We apply our results to a MoE generated by a quantum neural network.
Melchor Hernandez, A., Pastorello, D., De Palma, G. (2026). Mean-field limit from general mixtures of experts to quantum neural networks. LETTERS IN MATHEMATICAL PHYSICS, 116, 1-23 [10.1007/s11005-026-02065-9].
Mean-field limit from general mixtures of experts to quantum neural networks
Melchor Hernandez Anderson
;Pastorello Davide;De Palma Giacomo
2026
Abstract
In this work, we study the asymptotic behavior of mixture of experts (MoE) trained via gradient flow on supervised learning problems. Our main result establishes the propagation of chaos for a MoE as the number of experts diverges. We demonstrate that the corresponding empirical measure of their parameters is close to a probability measure that solves a nonlinear continuity equation, and we provide an explicit convergence rate that depends solely on the number of experts and on the dimensionality of the parameter space. We apply our results to a MoE generated by a quantum neural network.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.



