Statistical Advantages of Perturbing Cosine Router in Sparse Mixture of Experts
CoRR(2024)
摘要
The cosine router in sparse Mixture of Experts (MoE) has recently emerged as
an attractive alternative to the conventional linear router. Indeed, the cosine
router demonstrates favorable performance in image and language tasks and
exhibits better ability to mitigate the representation collapse issue, which
often leads to parameter redundancy and limited representation potentials.
Despite its empirical success, a comprehensive analysis of the cosine router in
sparse MoE has been lacking. Considering the least square estimation of the
cosine routing sparse MoE, we demonstrate that due to the intrinsic interaction
of the model parameters in the cosine router via some partial differential
equations, regardless of the structures of the experts, the estimation rates of
experts and model parameters can be as slow as 𝒪(1/log^τ(n))
where τ > 0 is some constant and n is the sample size. Surprisingly,
these pessimistic non-polynomial convergence rates can be circumvented by the
widely used technique in practice to stabilize the cosine router – simply
adding noises to the 𝕃_2 norms in the cosine router, which we
refer to as perturbed cosine router. Under the strongly identifiable
settings of the expert functions, we prove that the estimation rates for both
the experts and model parameters under the perturbed cosine routing sparse MoE
are significantly improved to polynomial rates. Finally, we conduct extensive
simulation studies in both synthetic and real data settings to empirically
validate our theoretical results.
更多查看译文
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要