Amazon Web Services introduced SWE-PolyBench, a multi-language benchmark for evaluating AI coding assistants.The benchmark aims to address limitations in current evaluation frameworks and assess AI agents in navigating complex codebases.SWE-PolyBench contains over 2,000 coding challenges across Java, JavaScript, TypeScript, and Python.It offers more diverse tasks and programming language support compared to existing benchmarks like SWE-Bench.The new benchmark introduces advanced evaluation metrics beyond pass rate, including file-level localization and CST node-level retrieval.Python remains the strongest language for AI agents, while performance declines with increased task complexity.Different agents exhibit varying strengths in bug fixing, feature requests, and code refactoring tasks.Clear issue descriptions significantly impact success rates for AI coding agents.SWE-PolyBench is crucial for evaluating AI coding assistants as they transition from experimental to production environments.The benchmark is publicly available for enterprise environments, providing valuable insights for real-world development scenarios.