Vanilla Regression VS Robust Regression

Vanilla Regression VS Robust Regression

Regression is one of the most widely used algorithms for forecasting. Regression is the first thing you'd learn in the ML world. For small problem statements, the basic versions of regression will work totally fine, but there are some cases where you can't use the vanilla regression technique. Having worked with regression, you might have noticed its sensitivity towards outliers. This comes from the least squares method that it uses to calculate the error. Errors with higher magnitude when squared, create an imbalance.

To address the sensitivity of the vanilla regression algorithm, a new technique evolved with emerging use cases of regression. In this article, we'll talk about two approaches: M - Estimation & R - Estimation.

M - Estimation

The idea here is to mimic the least-squares function when the residuals are near 0 and consider absolute values for larger residuals. Let's have a look at Huber's Dispersion Function to understand M - Estimation.

No alt text provided for this image

Here c is supposed to be the tolerance. Empirically, it's found that robust regression works the best when 1 < c < 2.

R - Estimation

Here every squared error is multiplied by some weight. This works like a ranking system to minimize the effect of larger residuals.

No alt text provided for this image
No alt text provided for this image

Where "a" is the weight used to create the rank effect.

The above are basic explanations for the two most used techniques to handle outliers.

要查看或添加评论,请登录

Rithwik Chhugani的更多文章

  • Why use Spark?

    Why use Spark?

    The very first reason you'll find on the internet is to work with Big Data, but why not use Pandas, Hadoop, or Dask?…

  • Tools for Smart/Lazy Data Scientists (ft. LazyPredict)

    Tools for Smart/Lazy Data Scientists (ft. LazyPredict)

    Being a data scientist you don't necessarily need to write tons and tons of code to see the performance of your models.…

  • Likelihood VS Probability

    Likelihood VS Probability

    It may look simple, but it's capable to create head-scratching situations at times. Let's understand in a few words…

  • Popular CNN Architectures

    Popular CNN Architectures

    Every now and then researchers try to fine-tune their existing model or come up with new architectures to win the…

  • Types of Hyperparameter Tuning

    Types of Hyperparameter Tuning

    What is hyperparameter tuning? Hyperparameter tuning is an extra step to make sure that your model is using the right…

社区洞察

其他会员也浏览了