I currently am working with a set of biological data and, at this time, have no dependent variable that I am trying to predict with these variables. However, I am wondering if there is a way to quantify the significance of these variables within the data set (i.e., how much information/variance they contribute to the data set) even though I do not have a variable I am trying to predict. I found methodology using a principal component analysis to determine a sub-group of variables that contribute the most information to the data set (Olliffe, 1972 - Method B4; King 1999 - Method B4). However, I have no way to quantify their significance. Any suggestions? Thank you in advance!
2026-03-29 14:18:16.1774793896
Quantifying Significance of Variables
53 Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in STATISTICS
- Given is $2$ dimensional random variable $(X,Y)$ with table. Determine the correlation between $X$ and $Y$
- Statistics based on empirical distribution
- Given $U,V \sim R(0,1)$. Determine covariance between $X = UV$ and $V$
- Fisher information of sufficient statistic
- Solving Equation with Euler's Number
- derive the expectation of exponential function $e^{-\left\Vert \mathbf{x} - V\mathbf{x}+\mathbf{a}\right\Vert^2}$ or its upper bound
- Determine the marginal distributions of $(T_1, T_2)$
- KL divergence between two multivariate Bernoulli distribution
- Given random variables $(T_1,T_2)$. Show that $T_1$ and $T_2$ are independent and exponentially distributed if..
- Probability of tossing marbles,covariance
Related Questions in DATA-ANALYSIS
- Calculating an average of other averages
- When extrapolating for projections, how do you know which function-form to use?
- Log base e versus log base 10 for semilog plot, how do I recognize the type of curve my data fits?
- Book Recommendation for Analyzing Experimental Results
- MLE of non-numerical variables?
- How to tell when a data series is a normal distribution
- Counting occurrences of elements from data stream
- Based on the coinmarket cap api, how do they calculate the bitcoin percent change
- Statistical analysis of study with categorical and numerical variables
- If I divide one column of a matrix by another, do I get new "information"?
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
Just some random thoughts:
Thought 1
Well, since you mentioned information or variance contributed to the data set $D=\{x_1,\ldots,x_n\}$ with features $F=(y_1,\ldots,y_m)$, you can literally just calculate it. For instance: $$ I_{1,D}(k) = E(D) - E(D_k) $$ where $x_i\in\mathbb{R}^m$, $E(D)$ is the joint entropy estimated from the full dataset, and $E(D_k)$ is the entropy of the dataset with feature $k$ removed. So this is the information entropy of the data with the feature minus it without the feature.
However, high dimensional entropies can be hard to compute. Another idea is to use the variance, as you mentioned. Let $C(D)$ be the covariance matrix of $D$. Then similar to above: $$ I_{2,D}(k) = ||C(D)||_p - ||C(D_k)||_p $$ where $p\in\{1,2,\infty,F\}$ (for instance) defines a matrix norm. This measures the difference between the covariance in the whole dataset and that of the dataset without the $k$th feature. One could also use the determinant instead of a matrix norm (which may give it greater connections to the entropy!)
Thought 2
Another idea is just to measure the total mutual information of each variable to each other one, and invert it. The idea would be that the most informative variables would be the ones with the least information shared with all the other variables For instance, $$ I_{3,D}(k) = \left[ \sum_i \mathcal{I}(y_k,y_i) \right]^{-1} $$ where $\mathcal{I}(y_k,y_i)$ is mutual information, or you could take the total information in the variable, and remove all the mutual informations between it and all the other variables: $$ I_{4,D}(k) = \exp\left(mH(y_k) - \sum_i \mathcal{I}(y_k,y_i)\right) $$ so you take the information entropy of feature $k$, $H(y_k)$, and subtract all the information shared between $y_k$ and the other features. Note that it can be rewritten as $$ I_{4,D}(k) = \exp\left( H(y_k) - \sum_{i\ne k} H(y_i) + H(y_k,y_i) \right) $$ so we take the information inside the feature, minus (1) the information in the other features and (2) the joint information held by the feature pairs. The exp is just to make the measure positive.
These are all measuring the "importance" of the variable by checking how non-redundant they are with the other variables.
If you want a "variance" measure rather than an information-theoretic one, you could use: $$ I_{5,D}(k) = \exp\left( 2\mathbb{V}[y_k] - \sum_{i} \text{Cov}[y_i,y_k] \right) $$
Thought 3
Also, here's something you could also consider doing with PCA. Let $v_i$ be the principal components with eigenvalues $\tilde{\lambda}_i$. I'll assume $v_i$ is normalized and let $\lambda_j=\tilde{\lambda}_j\left[\sum_k \lambda_k\right]^{-1}$ Then $v_i(k)$ is the $i$th eigenvector's $k$th component, i.e. the "importance" of feature $k$ in the PCA axis $i$. Notice that $\lambda_j$ is the variance explained by $v_j$.
Then, define $$ I_{6,D}(k) = \sum_i |v_i(k)| \lambda_i^p $$ for e.g. $p=1$ or $2$. Essentially, each PCA axis has an importance given by $\lambda_i$. So we take the contribution of feature $k$ to axis $i$, and weight it by the importance of that axis.