Defense MEDIUM
Defending MoE LLMs against Harmful Fine-Tuning via Safety Routing Alignment
Jaehan Kim, Minkyoo Song, Seungwon Shin +1 more
Recent large language models (LLMs) have increasingly adopted the Mixture-of-Experts (MoE) architecture for efficiency. MoE-based LLMs heavily depend...
AI Threat Alert