Skip to content

Comparing the use of original data vs PCA in multiple regression. Analysis also showcases how principal components can be transformed back to their original vector space after model fitting for descriptive purposes

Notifications You must be signed in to change notification settings

a-memme/predicting_crime_pca

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

22 Commits
 
 
 
 

Repository files navigation

Principal Component Analysis (PCA) and Predicting Crime Rates

Showcasing how PCA can be used as an effective dimensionality technique prior to model fitting, and how principal components can be transformed back to their original space for better interpretability

PCA

Principal Component Analysis is a dimensionality reduction method that becomes particularly useful when dealing with large datasets containing a wide array of dimensions. Through performing a linear transformation of the variables, PCA is effective in removing multicolinearity of the predictors, and through its ranking system, allows the analyst to further reduce the effects of randomness by concentrating on the top (components containing the most variance) principal components - i.e an effective technique for dimensionality reduction & simplifying the model.

Its main limitation lies in the fact that the response variable is never included in the calculation, and thus, no information provided by the response data is considered. The choosing of "important" factors in PCA has a bias toward variance present in the components themselves - i.e dimensions that have a larger variance are usually better predictors or differentiators, however, this in not ALWAYS the case. Thus, this is where the importance of informed trial and error - i.e the comparative evaluation and testing of several models - comes into play.

Analysis

Purpose

The following analysis aims to

  • Show how PCA can be used an effective dimensionality reduction technique
  • Showcase how principal components (new factors created by PCA) can be transformed back to their original vector space for easier interpretation/use after the model is fit.
    • (especially useful when the purpose of the analysis is predominately descriptive)
  • Explore limitations of PCA specifically present in this analysis.

Method

  • Multiple regression models are fit to the popular uscrime dataset using multiple dimensions to predict crime rate
  • PCA is applied to the predictor variables and models are fit to the newly created pricipal components.
  • Regression models using the original dataset vs PCA set are evaluated using 5-fold cross validation.
  • Principal components are then transformed back to their original vector space and models chosen via evaluation are tested and compared for performance.

Results

See pca_crime.R for details

Model Evaluation

Multiple Regression - Original Dataset
  • The model using only predictor variables deemed as significant (p < 0.05) or close to significant (p < 0.1) yielded the best performance in cross validation: image
Multiple Regression - PCA
  • Using the prcomp() function in R, the predictor variables in the original dataset are isolated, and pca is applied.
  • Once the principal components are merged back onto the respective response data, multiple regression models are fit to several different iterations of the data, based on model significance and/or scree plot inference (see below)
  • The model that performed best, similarily, was that that used only principal components deemded as significant in the original model: image

Expressing PCA in Original Terms (Reversing Linear Transformation)

  • The transformed coefficients are first multiplied by the matrix of eigenvectors to reverse the original rotation(s):
#Slice the coefficients to exclude the intercept, and multiply coefficients by eigenvectors (i.e rotations)
b_coeffs <- model$coefficients[2:5]
#reverse the rotation by multiplying coefficients by eigenvectors
c_scaled <- (b_coeffs %*% t(pca$rotation[, c(1,2,4,5)]))
  • The formula below represents how to transform standardized predictors back to their original form, where βj is the scaled regression coefficient of the jth predictor and β0 is the scaled intercept. In this step, we simply need to plug in the scaled coefficients, mean and standard deviation values after the rotation in the previous step (See point below for details):

    image

  • Finally, the same model created using the principal components derived from PCA, can be expressed in raw-value terms in the form of y = mx + b:

#Store the stdev and mean of the original data into variables -- the prcomp() function outputs us these values for each variable
#in the $scale (stdev) and $center (mean) dimensions
sigma <- pca1$scale
mu <- pca1$center

#Using the appropriate formula, find the unscaled values of An by dividing the scaled coefficient values by the stdev (reversing the scaling)
#solve for b
unscaled_coeffs <- c_scaled / sigma
unscaled_intercept <- model$coefficients[1] - sum(c_scaled * (mu/sigma))

#Final equation - in the style y= ax + b
prediction <- (as.matrix(crime_ivs) %*% unscaled_coeffs[1, ]) + unscaled_intercept

Discussion

PCA

  • The above analysis displays how PCA can be used as a dimensionality reduction technique prior to building a predictive model.
  • In our case, models using the original data vs the transformed data via PCA elicit very similar results, only seeing very slightly better performance in MAE and RMSE in the model trained with PCA data.

Original Vector Space

  • Transforming principal components back to their original vector space allows for:
    • presentation of the model that is compatible with the raw data but still represents the dimensionality of the Principal Component Analysis
    • aids significantly with model interpretation for descriptive analysis
    • provides more efficient functionality when considering ETL/automated processes.

Limitations

  • As mentioned in the introduction, the principal components chosen via PCA may not always lead directly to the most relevant metrics to the response variable. Sometimes, although a component may have greater variance, it may be a poorer predictor of the response variable. See the example of the 2D data below:

image

  • In our case, the scree plot below suggests that only the first 2-4 principal components should be used in the model based solely on the amount of variance present in each. As witnessed via cross-validation, the best performing regression model using principal components was the one that simply used all components deemed as significant from the original model - i.e the one that contained lower ranked principal components (such as PC12) which innately have less variance, and skip some higher ranked ones which innately have more (such as PC3).
#Visualize - scree plot to employ elbow method
screeplot(pca1, type='lines')

image

About

Comparing the use of original data vs PCA in multiple regression. Analysis also showcases how principal components can be transformed back to their original vector space after model fitting for descriptive purposes

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages