DyMoE: Continual Learning Dynamic MoE with Token-Level Guidance
Dynamic mixture-of-experts (MoE); Incrementing MoE for model expansion; Token assignment guidance for mitigating routing-drift & forgetting; continual learning of large vision-language models.