The popularity of Least Squares stems from its simplicity and its alignment with key statistical principles.The Least Squares approach is favored for its mathematical clarity, providing closed-form solutions for optimization.However, its reliability diminishes when data deviates from theoretical assumptions, like outliers in the distribution.Least Squares is especially useful in Linear Regression, offering optimal coefficients estimation through OLS.The conventional mean and median in statistics are directly linked to L2 and L1 losses, respectively.Under the Gauss-Markov theorem, the OLS estimator is acclaimed as the Best Linear Unbiased Estimator.OLS minimizes variance and maintains unbiasedness, outperforming other linear estimators.Least Squares is also equivalent to Maximum Likelihood Estimation in normal error scenarios.However, the reliance on normal errors makes Least Squares less effective in the presence of outliers.To address this limitation, robust loss functions like Huber or Tukey are recommended for outlier resilience.