<ul data-eligibleForWebStory="true">Randomization is being explored to boost adversarial robustness in machine learning models, with a focus on multiclass classification.Current theoretical analysis has mainly concentrated on binary classification, leaving gaps in understanding multiclass scenarios.A study draws from graph theory to analyze how randomization impacts adversarial risk minimization in multiclass settings.The analysis centers on discrete data distributions, mapping adversarial risk minimization to set packing problems.Three structural conditions on the data distribution's support are identified as crucial for randomization to enhance robustness.Switching from deterministic to randomized solutions in certain data distributions notably decreases optimal adversarial risk.The research underscores the significant role of randomization in fortifying multiclass classification against adversarial attacks.