Akaike information criterion

Akaike information criterion :

Akaike information criterion, also known as AIC, is a statistical measure used to evaluate the quality of a model by comparing the goodness of fit of the model with the number of parameters in the model. The AIC is calculated by adding the residual sum of squares (RSS) and twice the number of parameters in the model. The AIC is often used in model selection, where the goal is to choose the model with the smallest AIC value.
A simple example of using AIC in model selection can be seen in linear regression. Suppose we have a dataset with two predictor variables, x1 and x2, and one response variable, y. We want to use linear regression to model the relationship between the predictor variables and the response variable. We can fit two different models: a model with only x1 as a predictor, and a model with both x1 and x2 as predictors.
For the model with only x1 as a predictor, the model can be represented as y = β0 + β1×1. For the model with both x1 and x2 as predictors, the model can be represented as y = β0 + β1×1 + β2×2.
In order to calculate the AIC for each model, we first need to calculate the residual sum of squares (RSS) for each model. The RSS is calculated by summing the squared differences between the observed y values and the predicted y values for each observation.
For the model with only x1 as a predictor, the RSS is calculated as:
RSS1 = Σ(y – ŷ)2
where y is the observed y value for each observation, and ŷ is the predicted y value for each observation.
For the model with both x1 and x2 as predictors, the RSS is calculated as:
RSS2 = Σ(y – ŷ)2
where y is the observed y value for each observation, and ŷ is the predicted y value for each observation.
Once the RSS for each model is calculated, we can calculate the AIC for each model by adding the RSS and twice the number of parameters in the model. For the model with only x1 as a predictor, the AIC is calculated as:
AIC1 = RSS1 + 2(1)
where the 1 in the parentheses represents the number of parameters in the model (i.e., β0 and β1).
For the model with both x1 and x2 as predictors, the AIC is calculated as:
AIC2 = RSS2 + 2(2)
where the 2 in the parentheses represents the number of parameters in the model (i.e., β0, β1, and β2).
In this example, we can choose the model with the smallest AIC value, which would be the model with the best goodness of fit and the least number of parameters. In this case, the model with only x1 as a predictor would be chosen because it has a smaller AIC value than the model with both x1 and x2 as predictors.