Suppose we have the problem $$ \min_{q \in \mathcal{P}_n} \|f - q\|_{L^2(w)}^2 $$ where $\mathcal{P}_n$ is the space of polynomials of degree $n$, $w$ is some weight function (measure with continuous density), and $\|g\|_{L^2(w)}^2 = \int_a^b (g(x))^2 w(dx)$. I want to clarify the connection between this case and its finite dimensional analog $$ \min_{\beta \in \mathbb{R}^{n}} \|y - X\beta|_2^2. $$ In finite dimensions, we set the gradient equal to zero and recover the normal equations $$ 2X^T X \beta - 2X^T \beta = 0. $$ How can I recover the normal equations in the infinite dimensional case above? I'm at \begin{align} \min_{q \in \mathcal{P}_n} \|f - q\|_{L^2(w)}^2 &= \min_{q \in \mathcal{P}_n} \int_a^b [f(x)^2 - 2q(x)f(x) + q(x)^2] w(dx) \\ &= \min_{q \in \mathcal{P}_n} \int_a^b [- 2q(x)f(x) + q(x)^2] w(dx) \end{align} but I don't know how to differentiate wrt $q(x)$ (and am not sure that this is even the right step).
2026-02-23 06:18:20.1771827500
polynomial least squares derivation: normal equations
171 Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in POLYNOMIALS
- Alternate basis for a subspace of $\mathcal P_3(\mathbb R)$?
- Integral Domain and Degree of Polynomials in $R[X]$
- Can $P^3 - Q^2$ have degree 1?
- System of equations with different exponents
- Can we find integers $x$ and $y$ such that $f,g,h$ are strictely positive integers
- Dividing a polynomial
- polynomial remainder theorem proof, is it legit?
- Polyomial function over ring GF(3)
- If $P$ is a prime ideal of $R[x;\delta]$ such as $P\cap R=\{0\}$, is $P(Q[x;\delta])$ also prime?
- $x^{2}(x−1)^{2}(x^2+1)+y^2$ is irreducible over $\mathbb{C}[x,y].$
Related Questions in LEAST-SQUARES
- Is the calculated solution, if it exists, unique?
- Statistics - regression, calculating variance
- Dealing with a large Kronecker product in Matlab
- How does the probabilistic interpretation of least squares for linear regression works?
- Optimizing a cost function - Matrix
- Given matrix $Q$ and vector $s$, find a vector $w$ that minimizes $\| Qw-s \|^2$
- Defects of Least square regression in some textbooks
- What is the essence of Least Square Regression?
- Alternative to finite differences for numerical computation of the Hessian of noisy function
- Covariance of least squares parameter?
Related Questions in WEIGHTED-LEAST-SQUARES
- Weighted Least Squares Without Intercept
- Understanding Galerkin method of weighted residuals
- Is there a least squares estimator for correlated, non-constant variance errors?
- How can I get the gradient of the normal equation for weighted linear regression?
- Minimizing a multivariable function with different vector length?
- Weighted Least Squares - Categorical Data vs. Numerical Data
- Matrix notation for weighted sum of squares
- How to solve conditional least square?
- Nearest (with respect to weights) symmetric positive semidefinite matrix
- Understanding an identity for least squares regression line gradient
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
The Normal Equations actually hold in any Hilbert space as long as the subspace you are estimating in is of finite (co-)dimension and here $V=\mathcal P_n$ has $\dim n+1$.
To see this, observe by the Projection Theorem that the optimal vector $q=\sum a_j p_j, p_j \in V$ satisfies \begin{align*} \left\langle x -\sum a_j p_j | p_i \right\rangle=0. \end{align*} Use this to derive expressions for the $a_j$ and you end up with an infinite-dimensional generalization of the normal equations (i.e. your result will reduce to the matrix form if your Hilbert space itself is finite-dimensional).
Luenberger's classic optimization by vector space methods has a beautiful exposition of this topic in chapter 3, see also connections to projection operators and pseudoinverses.
Edit: in your case \begin{align*} \langle x| y\rangle = \int xy d\mu \end{align*} and $w(dx)=d\mu$.
Edit2: Your attempt to differentiate wrt $q$ is essentially the same as in the calculus of variations, you could probably get somewhere with Euler-Lagrange as well if you still want to try that route. This connects to the idea that there often are (at least) two ways to solve an (easy) optimization problem: the direct route via differential techniques, here Euler-Lagrange, or the geometric route, here the projection theorem, using orthogonality.