The final step to deal with outliers is to evaluate the effect and significance of the outliers on your ML models. This can help you decide if removing, replacing, or scaling outliers is beneficial or detrimental for your specific ML task, and how to refine your data cleaning and preprocessing strategies. Descriptive statistics or visualizations can be used to compare the summary and distribution of your data before and after dealing with outliers. This way, you can observe changes in the mean, median, variance, range, skewness, or kurtosis. Additionally, inferential statistics or hypothesis tests can be used to compare the significance and confidence of your data before and after dealing with outliers. This can help validate and justify your decisions about outliers regarding the p-value, t-test, ANOVA, or chi-square test. Lastly, machine learning metrics or validation techniques can be used to compare the performance and accuracy of your ML models before and after dealing with outliers. This way, you can measure and optimize your ML outcomes and objectives such as accuracy, precision, recall, F1-score, MSE, R2, or cross-validation. Python libraries like pandas, matplotlib, seaborn, scipy, statsmodels or sklearn can be used to evaluate outliers with these methods depending on the type and goal of your ML task.