A research analyzing the impact of feature scaling on Machine Learning found distinct effects on various algorithms and datasets in classification and regression tasks.
The study evaluated 12 scaling techniques across 14 ML algorithms and 16 datasets, showing differences in predictive performance metrics and computational costs.
Ensemble methods like Random Forest and XGBoost demonstrated consistent performance regardless of scaling techniques, while Logistic Regression, SVMs, TabNet, and MLPs showed significant performance variations depending on the scaler used.
The research provides valuable insights for practitioners by emphasizing the importance of choosing appropriate feature scaling techniques based on specific machine learning models.