In Gilbert Strang's 'Introduction to Applied Mathematics', in chapter 2.5 (Least Squares Estimation and the Kalman Filter), in proof 2I, he talks about 'minimizing the covariance matrix'. I can't determine what the criteria is for minimizing a matrix. Does anyone have insight to lend?
2026-03-31 22:50:11.1774997411
What does it mean to minimize a matrix
1.5k Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in LINEAR-ALGEBRA
- An underdetermined system derived for rotated coordinate system
- How to prove the following equality with matrix norm?
- Alternate basis for a subspace of $\mathcal P_3(\mathbb R)$?
- Why the derivative of $T(\gamma(s))$ is $T$ if this composition is not a linear transformation?
- Why is necessary ask $F$ to be infinite in order to obtain: $ f(v)=0$ for all $ f\in V^* \implies v=0 $
- I don't understand this $\left(\left[T\right]^B_C\right)^{-1}=\left[T^{-1}\right]^C_B$
- Summation in subsets
- $C=AB-BA$. If $CA=AC$, then $C$ is not invertible.
- Basis of span in $R^4$
- Prove if A is regular skew symmetric, I+A is regular (with obstacles)
Related Questions in COVARIANCE
- Let $X, Y$ be random variables. Then: $1.$ If $X, Y$ are independent and ...
- Correct formula for calculation covariances
- How do I calculate if 2 stocks are negatively correlated?
- Change order of eigenvalues and correspoding eigenvector
- Compute the variance of $S = \sum\limits_{i = 1}^N X_i$, what did I do wrong?
- Bounding $\text{Var}[X+Y]$ as a function of $\text{Var}[X]+\text{Var}[Y]$
- covariance matrix for two vector-valued time series
- Calculating the Mean and Autocovariance Function of a Piecewise Time Series
- Find the covariance of a brownian motion.
- Autocovariance of a Sinusodial Time Series
Related Questions in LEAST-SQUARES
- Is the calculated solution, if it exists, unique?
- Statistics - regression, calculating variance
- Dealing with a large Kronecker product in Matlab
- How does the probabilistic interpretation of least squares for linear regression works?
- Optimizing a cost function - Matrix
- Given matrix $Q$ and vector $s$, find a vector $w$ that minimizes $\| Qw-s \|^2$
- Defects of Least square regression in some textbooks
- What is the essence of Least Square Regression?
- Alternative to finite differences for numerical computation of the Hessian of noisy function
- Covariance of least squares parameter?
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
Strang’s style is quite informal and I don’t particularly care for it (though it apparently appeals to some people). In this case, a more formal presentation would have certainly spelled out what he meant. Looking over his book's proof of 2I, I would say that Strang most likely means minimization of the matrix $P$ in the spectral norm over all matrices $L$ satisfying $LA=I$. Here’s why I think so:
In Strang’s notation, $V$ is the covariance matrix which is symmetric positive semi-definite. The matrix he says he wants to minimize has the form $P=LVL^T$, which (following some context-specific simplifications) he then breaks down as $P= L_0V{L_0}^T+ (L-L_0)V(L-L_0)^T$. $P$ is clearly a real symmetric (hence normal) positive semi-definite matrix which implies that $$\left\| P \right\|_2 = \sigma _{\max } \left( P \right) = \lambda_{\max } \left( P \right) = \mathop {\max }\limits_{\left\| x \right\|_2 = 1} x^T Px.$$
Note that $L$ is the only free parameter in Strang’s expression for $P$. If we assume that Strang is attempting to minimize the spectral norm of $P$ over all $L$ satisfying $LA=I$, then this is accomplished by finding $L$ such that $\mathop {\min }\limits_{\{ L|LA = I\} } \left\| P \right\|_2 = \mathop {\min }\limits_{\{ L|LA = I\} } \mathop {\max }\limits_{\left\| x \right\|_2 = 1} x^T Px$. That is, $$\mathop {\min }\limits_{\{ L|LA = I\} } \left\| P \right\|_2 = \mathop {\min }\limits_{\{ L|LA = I\} } \mathop {\max }\limits_{\left\| x \right\|_2 = 1} \left\{ {x^T L_0 VL_0^T x + x^T (L - L_0 )V(L - L_0 )^T x} \right\}$$
The second quadratic form, the only component involving $L$, is clearly bounded below by $0$ and this bound is achieved by taking $L=L_0$. This gives $P = L_0 VL_0^T$ which is what Strang wanted to show.
As an aside, in looking over his proof, I noticed that Strang has a minor typo in one of the intermediate steps. He writes$$(L - L_0 )VL_0 ^T = (L - L_0 )VV^{ - 1} A(A^T V^{ - 1} A)^{ - 1}$$
when he should have written
$$(L - L_0 )VL_0 ^T = (L - L_0 )VV^{ - 1} A\left\{ {(A^T V^{ - 1} A)^{ - 1} } \right\}^T.$$