If $A$ is a positive definite matrix, then the solution to the minimization problem $(1/2)x^TAx - b^Tx$ is given by $A^{-1}b$. I'm interested in the generalization of this to a Hilbert space. What conditions on $A$ should be required for this to be true in a Hilbert space? Also is there a reference that includes this? I searched Google books, but couldn't find it.
2026-04-07 05:01:54.1775538114
Quadratic minimization in a Hilbert space
1.6k Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in FUNCTIONAL-ANALYSIS
- On sufficient condition for pre-compactness "in measure"(i.e. in Young measure space)
- Why is necessary ask $F$ to be infinite in order to obtain: $ f(v)=0$ for all $ f\in V^* \implies v=0 $
- Prove or disprove the following inequality
- Unbounded linear operator, projection from graph not open
- $\| (I-T)^{-1}|_{\ker(I-T)^\perp} \| \geq 1$ for all compact operator $T$ in an infinite dimensional Hilbert space
- Elementary question on continuity and locally square integrability of a function
- Bijection between $\Delta(A)$ and $\mathrm{Max}(A)$
- Exercise 1.105 of Megginson's "An Introduction to Banach Space Theory"
- Reference request for a lemma on the expected value of Hermitian polynomials of Gaussian random variables.
- If $A$ generates the $C_0$-semigroup $\{T_t;t\ge0\}$, then $Au=f \Rightarrow u=-\int_0^\infty T_t f dt$?
Related Questions in OPTIMIZATION
- Optimization - If the sum of objective functions are similar, will sum of argmax's be similar
- optimization with strict inequality of variables
- Gradient of Cost Function To Find Matrix Factorization
- Calculation of distance of a point from a curve
- Find all local maxima and minima of $x^2+y^2$ subject to the constraint $x^2+2y=6$. Does $x^2+y^2$ have a global max/min on the same constraint?
- What does it mean to dualize a constraint in the context of Lagrangian relaxation?
- Modified conjugate gradient method to minimise quadratic functional restricted to positive solutions
- Building the model for a Linear Programming Problem
- Maximize the function
- Transform LMI problem into different SDP form
Related Questions in HILBERT-SPACES
- $\| (I-T)^{-1}|_{\ker(I-T)^\perp} \| \geq 1$ for all compact operator $T$ in an infinite dimensional Hilbert space
- hyponormal operators
- a positive matrix of operators
- If $S=(S_1,S_2)$ hyponormal, why $S_1$ and $S_2$ are hyponormal?
- Is the cartesian product of two Hilbert spaces a Hilbert space?
- Show that $ Tf $ is continuous and measurable on a Hilbert space $H=L_2((0,\infty))$
- Kernel functions for vectors in discrete spaces
- The space $D(A^\infty)$
- Show that $Tf$ is well-defined and is continious
- construction of a sequence in a complex Hilbert space which fulfills some specific properties
Related Questions in CALCULUS-OF-VARIATIONS
- On sufficient condition for pre-compactness "in measure"(i.e. in Young measure space)
- Weak formulation of Robin boundary condition problem
- Why is the index of a harmonic map finite?
- Variational Formulation - inhomogeneous Neumann boundary
- Relationship between Training Neural Networks and Calculus of Variations
- How to prove a Minimal Surface minimizes Surface Tension
- Derive the Euler–Lagrange equation for a functional a single variable with higher derivatives.
- Does the covariant derivative commute with the variational derivative?
- Derivative of a functional w.r.t. a single point?
- calculus of variations with double integral textbook?
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
What you need is to look at some texts on functional analysis, especially about "Direct Methods" in "Calculus of Variations". I'll give a quick primer here to flesh out what Rahul Narain said in the comments, but it will be by no means complete.
First let us review how the minimization problem is solved in finite dimensional spaces. We will consider a finite dimensional vector space $V$ with an inner product $\langle,\rangle$. Let $A$ be a self-adjoint linear operator $V\to V$, and $b$ some element of $V$. We consider the minimization problem with the non-linear functional $J(x) = \frac12 \langle x,Ax\rangle - \langle b,x\rangle$. There are two steps used in finding the minimizer.
Step 1) Existence of a minimizer. First we need to show that $J(x)$ is bounded from below. This is where we first use that $A$ has only non-negative eigenvalues. If $A$ were to have negative eigenvalues, let $v$ be its eigenvector, and look at $J(kv)$ for $k$ a scalar. Letting $k\nearrow\infty$ we see that $J(kv)\searrow -\infty$ and there cannot be an minimum. Next, if $A$ were to have a non-empty kernel, such that there exists some vector $v$ with the property $Av = 0$ and $\langle b,v\rangle \neq 0$, then again taking $J(kv)$ and sending $k$ to either $\pm\infty$, we will have that the functional $J$ is not bounded below.
So $A$ must only have positive eigenvalues. Now, let $\lambda$ be the smallest of $A$'s eigenvalues. Then we have $\langle x,Ax\rangle > \lambda \|x\|^2$. Using Cauchy-Schwarz inequality, we then have
$$ J(x) > \frac12 \lambda \|x\|^2 - \|b\|\|x\| $$
So we see that for $\|x\| > 2\|b\| / \lambda$, $J(x) > 0$. And by continuity, $J(x)$ is bounded below.
Now, let $j = \inf J < 0$. (The infimum is negative since by choosing $w$ a sufficiently small multiple of $b$, we see that $J(w) < 0$.) By the definition of infimum, there exists a sequence of vectors $(v_k)$ such that $J(v_k) \searrow j$. Since the infimum is less than 0, by our previous computations we know that the sequence $(v_k)$ must all satisfy $\|v_k\| \leq 2\|b\|/\lambda$. Now, this set is a bounded closed ball in a finite dimensional vector space. In particular, it is compact. Hence up to a subsequence $v_k$ must converge. Now using continuity of $J$, you have that the limit element $v_\infty$ of the sequence attains the minimum. That is $J(v_\infty) = j$.
Step 2) Computing the minimizer. Here we just use the Euler-Lagrange equations. At the minimizer, the functional is stationary. That is $(\frac{\delta}{\delta x}J)(v_\infty) = 0$. Taking the variation of the functional $J$, we get, using again that $A$ is self-adjoint, that
$$ \frac{\delta}{\delta x}J = 0 \iff Ax - b = 0 $$
Since $A$ only has positive eigenvalues, it is invertible, so we can solve $v_\infty = A^{-1}b$.
Formally, we can try to do the same thing in infinite dimension. But there are a few caveats. First is that it is not enough that $A$ only has positive eigenvalues. You need that $A$ is coercive in the following sense:
$$ \langle x, Ax\rangle > \lambda \|x\|^2$$
for some $\lambda$. In finite dimensions that $A$ is positive implies the existence of this $\lambda$. In infinite dimensions this is not true. Imagine the $\ell^2$ Hilbert space with standard inner product and the standard basis $(e_k)$. Let $Ae_k = \frac1k e_k$. Then $A$ has only positive eigenvalues. But by taking larger and larger $k$'s, we can make the eigenvalue approach 0 as closely as we want, and violate coercivity. Coercivity allows us to guarantee that the functional $J(x)$ has a bottom.
The second problem is that a closed bounded set in a Hilbert space is not necessarily compact. So we cannot conclude immediately that a minimizing sequence converges in the norm topology. However, using Banach-Alaoglu theorem, the minimizing sequence must have a convergent subsequence in the weak topology of your Hilbert space. So it is sufficient to show that your functional $J(x)$ is what is called "sequentially weakly lower-semicontinuous" to get the desired conclusion. (That condition just guarantees that for any weakly converging sequence $w_k\to_w w_\infty$, $\lim\inf J(w_k) \geq J(w_\infty)$.)
In our particular case, things can be made slightly easier. If we in addition assume that $A$ is a bounded linear operator. Then we can use $(x,y) = \langle x,Ay\rangle$ as an equivalent inner product on our Hilbert space. Then $A^{-1}b$ exists by the Riesz representation theorem, and our functional can be written as $J(x) = \frac12 (x,x) - (A^{-1}b, x)$. Now we run through our previous arguments again. Now $w_k\to_w w_\infty$, and therefore $(A^{-1}b,w_k) \to (A^{-1}b,w_\infty)$ strongly by the definition of weak limits. And we also use the property of weak limits that
$$ \lim\inf (w_k,w_k) \geq (w_\infty,w_\infty)$$
which gurantees lower semi-continuity as desired. So now we've shown that the minimizer exists, and computing the variation we see that the minimizer must satisfy $w_\infty = A^{-1}b$.
So to repeat, a sufficient condition is that $A$ is self-adjoint, bounded, and coercive.