Mixture-of-Experts models achieve performance and inference efficiency by activating only a subset of experts.Large-scale Mixture-of-Experts models face the limitation of storing all experts which leads to significant memory overhead.A pruning framework called EASY-EP is proposed, which utilizes domain-specific demonstrations to identify and retain the most relevant experts.EASY-EP can achieve comparable performance and higher throughput while reducing memory usage by half in the DeepSeek-R1 model.