I am wondering if there is a practical interpretation of a principal component analysis: Consider you have a data matrix $X\in\mathbb{R}^{N\times p}$ and you perform a principal component analysis where you typically receive certain directions $v_1,...,v_q$, $q<p$, in $\mathbb{R}^N$ that explain the most of the variance in the data. Is there an interpretation of these principal components in terms of the original components, i.e. the variables $x_1,...,x_p$ that constitute the model. Think e.g. of $x_i$ being certain "variables" of a human body such as weight, blood pressure etc. that should be used to predict expected life time. If one now performs a PCA as described a above, one recognizes that certain linear combinations of the columns of $X$ explain most of the variance. If one wants to reduce the model (i.e. reduce the $p$), which variables do you exclude given the information of the PCA?
2025-01-13 02:27:34.1736735254
Interpretation of PCA
227 Views Asked by JohnSmith https://math.techqa.club/user/johnsmith/detail At
1
There are 1 best solutions below
Related Questions in STATISTICS
- Finding the value of a constant given a probability density function
- How to find probability after finding the CDF of a max?
- Is the row space of a matrix (order n by m, m < n) of full column rank equal to $\mathbb{R}^m$?
- Rate of growth of expected value of maxima of iid random variables under minimal assumptions
- Lower bound for the cumulative distribution function on the left of the mean
- Hypergeometric Probability
- $\mathbb E[(\frac{X+1}{4}-\theta)^2]=?$
- What value of $\alpha$ makes $\sum_{i=0}^n (x_i-\alpha)^2$ minimum?
- The distribution of fourier coefficients of a Rademacher sequence
- small sample (1 point) MLE estimation
Related Questions in REGRESSION
- small sample (1 point) MLE estimation
- For linear regression: compute $\Theta T X$
- Nonlinear LS regression
- Least-squares fit of a nonlinear (polar) system
- Discrete version of continuous SIR model
- How to get the relative contribution of each variable in a difference that forms the numerator?
- Why is $R^2=\rho^2$
- Please show that $f(\beta_0,\beta_1)=\log(1+\operatorname{exp}(-y_1(\beta_0+\beta_1 x_1)))+\log(1+\operatorname{exp}(-y_2(\beta_0+\beta_1 x_2)))$
- standardised random variable least square regression $X$ against $Y$, $Y$ against $X$
- Interpreting OLS Regression Coefficients with High Multicolinearity
Related Questions in REGRESSION-ANALYSIS
- Rounding in the method of least squares for linear regression analysis?
- Sxx in linear regression
- Difference beween Residual and Error in Regression
- E(b1)=beta1 regression
- Linear regression with a given (non zero) intercept
- Variance of residuals from simple linear regression
- How to calculate parameters of a logarithmic approximation trendline?
- Interpretation of PCA
- The four assumptions on linear regression
- Multiple Regression Forecast
Related Questions in PRINCIPAL-COMPONENT-ANALYSIS
- PCA Interpretation
- Fitting a plane to points using SVD
- Principal Component Analysis Linear Transformation
- For PCA using the eigenvectors of the covariance matrix, what is the meaning of the eigenvalues?
- A beginner’s explanation for PCA on a multivariate time series
- How to solve the optimization problem of PCA?
- Principal Component Analysis Explanation
- PCA and linear combinations
- Derivation of PCA using Multivariable Calculus
- Does changing units affect PCA?
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Refuting the Anti-Cantor Cranks
- Find $E[XY|Y+Z=1 ]$
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- What are the Implications of having VΩ as a model for a theory?
- How do we know that the number $1$ is not equal to the number $-1$?
- Defining a Galois Field based on primitive element versus polynomial?
- Is computer science a branch of mathematics?
- Can't find the relationship between two columns of numbers. Please Help
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- A community project: prove (or disprove) that $\sum_{n\geq 1}\frac{\sin(2^n)}{n}$ is convergent
- Alternative way of expressing a quantied statement with "Some"
Popular # Hahtags
real-analysis
calculus
linear-algebra
probability
abstract-algebra
integration
sequences-and-series
combinatorics
general-topology
matrices
functional-analysis
complex-analysis
geometry
group-theory
algebra-precalculus
probability-theory
ordinary-differential-equations
limits
analysis
number-theory
measure-theory
elementary-number-theory
statistics
multivariable-calculus
functions
derivatives
discrete-mathematics
differential-geometry
inequality
trigonometry
Popular Questions
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- How to find mean and median from histogram
- Difference between "≈", "≃", and "≅"
- Easy way of memorizing values of sine, cosine, and tangent
- How to calculate the intersection of two planes?
- What does "∈" mean?
- If you roll a fair six sided die twice, what's the probability that you get the same number both times?
- Probability of getting exactly 2 heads in 3 coins tossed with order not important?
- Fourier transform for dummies
- Limit of $(1+ x/n)^n$ when $n$ tends to infinity
I assume that by "reducing the model," you mean to omit those measurements from the predictive model entirely (such that, for instance, you wouldn't even need to collect information about those variables in the future).
For this kind of goal, ordinary PCA is inappropriate, because each new variable is, as you said, a linear combination of the old variables.
If you want to stay within the realm of principal components analysis, then you should look at Sparse PCA. Sparse PCA essentially finds an approximation to the true PCA under the constraint that the principal components must be sparse.
However, there are two disadvantages here:
Thus, it sounds like what you really want is the lasso. In the realm of predictive modeling, the lasso would find a subset of predictor variables that are useful for predicting the response variable.