Standardized coefficient
In statistics, standardized coefficients, also called beta coefficients or beta weights, are the estimates resulting from a regression analysis where the underlying data have been standardized so that the variances of dependent and independent variables are equal to 1. Therefore, standardized coefficients are unitless and refer to how many standard deviations a dependent variable will change, per standard deviation increase in the predictor variable.
Usage
Standardization of the coefficient is usually done to answer the question of which of the independent variables have a greater effect on the dependent variable in a multiple regression analysis where the variables are measured in different units of measurement.It may also be considered a general measure of effect size, quantifying the "magnitude" of the effect of one variable on another.
For simple linear regression with orthogonal predictors, the standardized regression coefficient equals the correlation between the independent and dependent variables.
Implementation
A regression carried out on original variables produces unstandardized coefficients. A regression carried out on standardized variables produces standardized coefficients. Values for standardized and unstandardized coefficients can also be re-scaled to one another subsequent to either type of analysis.Suppose that is the regression coefficient resulting from a linear regression. The standardized coefficient simply results as, where and are the (estimated) standard deviations of and, respectively.
Sometimes, standardization is done only without respect to the standard deviation of the regressor.
Advantages and disadvantages
Standardized coefficients' advocates note that the coefficients are independent of the involved variables' units of measurement, which makes comparisons easy.Critics voice concerns that such a standardization can be very misleading.
Due to the re-scaling based on sample standard deviations, any effect apparent in the standardized coefficient may be due to confounding with the particularities of the involved data sample.
Also, the interpretation or meaning of a "one standard deviation change" in the regressor may vary markedly between non-normal distributions.