codingstreets
Search
Close this search box.

Assumption of Linear Regression in Machine Learning

code projected over woman
Photo by ThisIsEngineering on Pexels.com

This article is about the Linear Regression Assumption which summarizes the importance of assessing assumptions in linear regression analysis.

Also, take a look at Introduction to Linear Regression

Table of Contents

What is Linear Regression?

Linear regression is a statistical method used to analyze the relationship between two variables. It involves finding a line that best fits a set of data points so that we can make predictions about one variable based on the value of another. In other words, it helps us to estimate the value of one variable based on the known value of another. 

For example, if we wanted to predict someone’s salary based on their years of experience, we could use linear regression to find a line that best fits the data we have on salaries and years of experience, and then use that line to predict the salary of someone with a certain number of years of experience.

The line that we find using linear regression is called the regression line, and it represents the best estimate of the relationship between the two variables. The slope of the regression line tells us how much the value of the dependent variable changes for every one unit change in the independent variable. The intercept of the regression line tells us the value of the dependent variable when the independent variable is zero.

Linear regression can be used for both simple and multiple linear regression problems, depending on the number of independent variables involved.

What are the Assumptions of Linear Regression?

Linear regression makes certain assumptions about the data that need to be met for the results to be reliable. Here are the key assumptions of linear regression:

  1. Linearity: The relationship between the dependent variable and each independent variable should be linear, meaning that the change in the dependent variable is proportional to the change in the independent variable.
  1. Independence: The observations should be independent of each other, meaning that the value of another observation should not influence the value of one observation.
  1. Homoscedasticity: The variance of the errors should be constant across all levels of the independent variables. In other words, the spread of the residuals should be the same across the range of the independent variable.
  1. Normality: The residuals should be normally distributed, meaning that they should follow a bell-shaped curve around zero.
  1. No Multicollinearity: The independent variables should not be highly correlated with each other, as this can lead to problems with interpreting the coefficients and can make the results unreliable.

It is important to check these assumptions before running a linear regression model to ensure the results are valid and trustworthy.

How do you know if this assumption is correct in Homoscedasticity?

Homoscedasticity, also known as constant variance, is an important assumption of linear regression. We can check for homoscedasticity by examining the residuals (the differences between the predicted values and the actual values) plotted against the predicted values or the independent variable.

If the residuals are randomly scattered around zero and the spread of the residuals is consistent across the range of the predicted values or independent variable, then we can assume that the data has homoscedasticity. This means that the variability of the dependent variable is similar across all levels of the independent variable.

However, if the residuals are not randomly scattered around zero and/or the spread of the residuals changes as the predicted values or independent variable changes, then the data may have heteroscedasticity, which violates the assumption of homoscedasticity. In this case, the variability of the dependent variable is different at different levels of the independent variable.

One way to check for homoscedasticity is by creating a scatter plot of the residuals against the predicted values or the independent variable. If the scatter plot has a funnel shape, with the spread of the residuals increasing or decreasing as the predicted values or independent variable increases, then the data may have heteroscedasticity. On the other hand, if the scatter plot has a random pattern and the spread of the residuals is relatively constant across the range of the predicted values or independent variable, then the data may have homoscedasticity. 

Another way to check for homoscedasticity is by performing a statistical test, such as the Breusch-Pagan test or the White test. These tests can provide a formal test for the presence of heteroscedasticity in the data. If the test indicates the presence of heteroscedasticity, we may need to consider using alternative regression models or transformations to address the issue.

Common ways to fix Heteroscedasticity

Yes, there are several ways to address the issue of heteroscedasticity in linear regression. Here are three common methods:

  1. Weighted least squares regression: This method involves assigning more weight to observations with smaller residuals and less weight to observations with larger residuals. By doing so, we can give more emphasis to the observations that have less variability and are more reliable, and reduce the influence of observations that have more variability and are less reliable.

  2. Transformations: We can transform the dependent variable or the independent variable by taking the logarithm or square root, for example. This can help to reduce the variability of the data and make the relationship more linear. However, it is important to choose an appropriate transformation that does not distort the meaning of the data.

  3. Heteroscedasticity-consistent standard errors: This method involves using a different method of estimating the standard errors of the regression coefficients that takes into account the presence of heteroscedasticity. This can help to produce more accurate estimates of the standard errors and can also help to produce more accurate hypothesis tests and confidence intervals.

It is important to note that the choice of method for addressing heteroscedasticity depends on the specific characteristics of the data and the research question. Additionally, it is important to check that the method used has not introduced new problems or biases into the analysis.

Methods for determining assumption in Normality 

There are several methods to determine whether the assumption of normality is met for a given dataset. Here are a few commonly used methods:

  1. Visual inspection: One of the simplest ways to assess normality is to create a histogram of the data and visually inspect it for a bell-shaped curve. You can also create a probability plot or a Q-Q plot, which compares the distribution of the data to a theoretical normal distribution. If the points on the plot follow a straight line, the data is approximately normal.

An example of residuals that nearly follow a normal distribution is seen in the Q-Q plot below:

Normal Q-Q Plot - codingstreets
Source: AU

The Q-Q plot below, on the other hand, illustrates an example of when residuals obviously deviate from a straight diagonal line, indicating that they do not follow a normal distribution:

Normal Q Plot - codingstreets
Source: AU
  1. Shapiro-Wilk test: The Shapiro-Wilk test is a statistical test that can be used to assess normality. It tests the null hypothesis that the data is normally distributed. If the p-value is less than the significance level (usually 0.05), the null hypothesis is rejected and the data is considered non-normal.
  1. Kolmogorov-Smirnov test: The Kolmogorov-Smirnov test is another statistical test that can be used to assess normality. It tests the null hypothesis that the data comes from a normal distribution. If the p-value is less than the significance level, the null hypothesis is rejected and the data is considered non-normal.
  1. Skewness and kurtosis: Skewness and kurtosis are measures of the shape of a distribution. If the skewness is close to zero and the kurtosis is close to three (the values for a normal distribution), then the data is approximately normal.

It’s important to note that no method is perfect and each method has its own limitations. Therefore, it is recommended to use multiple methods to determine whether the assumption of normality is met. If the assumption is not met, it may be necessary to use methods to transform the data or use non-parametric tests that do not rely on the assumption of normality.

Common ways to fix Normality

When analyzing data using statistical methods, one assumption that is often made is that the data is normally distributed. If the data is not normally distributed, we may need to take steps to transform the data in order to meet this assumption. Here are three common ways to fix normality:

  1. Logarithmic or exponential transformation: If the data is skewed to the right (i.e., has a long tail on the right-hand side), we can try taking the logarithm or exponential of the data. This can help to make the data more symmetric and closer to a normal distribution.

  2. Box-Cox transformation: The Box-Cox transformation is a general method that can be used to transform data that is not normally distributed into a normal distribution. It involves taking the data to a power that is determined by a parameter, which is estimated from the data. This method is often used when it is not clear what type of transformation to use.

  3. Non-parametric methods: If the data is severely non-normal and cannot be transformed to meet the normality assumption, non-parametric methods such as the Wilcoxon rank-sum test or Kruskal-Wallis test can be used. These methods do not assume a normal distribution and can be used when the data is heavily skewed or has outliers.

It is important to note that the choice of method for fixing normality depends on the specific characteristics of the data and the research question. Additionally, it is important to check that the method used has not introduced new problems or biases into the analysis.

How to determine if this assumption is met in Independence

The assumption of independence is important in many statistical analyses, as violating it can lead to biased or unreliable results. Here are some ways to determine if this assumption is met:

  1. Study design: Independence is often assumed when data is collected using a randomized study design, such as a randomized controlled trial. In this case, the assignment of participants to treatment groups is done randomly, ensuring that any observed differences between groups are due to the treatment and not other factors.

  2. Data collection methods: The methods used to collect data can also impact independence. For example, if participants are allowed to influence each other in a study (e.g., through social interactions), independence may be violated. Similarly, if data is collected at different times or under different conditions, these factors may impact the observed outcomes.

  3. Correlation analysis: One way to check for independence is to perform a correlation analysis between pairs of variables. If there is a strong correlation between two variables, this suggests that they may not be independent, and may need to be adjusted for in any statistical analysis.

  4. Time series analysis: In some cases, data is collected over time, such as in a longitudinal study. In this case, it is important to consider the time-dependence of the data and adjust for any potential autocorrelation that may exist.

It is important to carefully consider the assumptions of independence when designing a study or analyzing data. Violations of independence can lead to biased results and incorrect conclusions, so it is important to take steps to ensure that the data is as independent as possible.

Conclusion

In conclusion, linear regression is a widely used statistical technique for modeling the relationship between a dependent variable and one or more independent variables. However, it is important to ensure that the assumptions of linear regression are met in order to obtain accurate and reliable results.

The key assumptions of linear regression include linearity, independence, homoscedasticity, and normality. Linearity assumes that the relationship between the dependent variable and the independent variable(s) is linear, and can be checked using scatter plots or correlation analysis. Independence assumes that the observations are independent of each other, and can be checked using study design or correlation analysis. Homoscedasticity assumes that the variance of the errors is constant across all levels of the independent variable(s), and can be checked using residual plots or statistical tests. Normality assumes that the errors are normally distributed, and can be checked using probability plots or statistical tests.

If any of these assumptions are violated, it can lead to biased or unreliable results. Therefore, it is important to carefully check for these assumptions before using linear regression for modeling or prediction. If the assumptions are not met, alternative modeling techniques may be required, or it may be necessary to transform the data to meet the assumptions.

In summary, understanding and assessing the assumptions of linear regression is crucial for obtaining accurate and reliable results, and researchers should take care to ensure that these assumptions are met before using this technique for analysis.

Recent Articles