I have read in a paper that for following function $$f(\textbf{x})=\frac{1}{a\|\textbf{x}\|^2+b}$$ we can have following concave surrogate function $$g(\textbf{x}|\textbf{x}_0)=-\frac{a\|\textbf{x}\|^2}{b^2}+2a\left(\frac{1}{b^2}-\frac{1}{(a\|\textbf{x}_0\|^2+b)^2}\right)\textbf{x}^T\textbf{x}_0+\frac{1}{(a\|\textbf{x}_0\|^2+b)}+\frac{2a\|\textbf{x}_0\|^2}{(a\|\textbf{x}_0\|^2+b)^2}-\frac{a\|\textbf{x}_0\|^2}{b^2}$$ where $a$ and $b$ are both positive constants. I will be very thankful if somebody could provide a method for obtaining the concave surrogate function. Any help in this regard will be much appreciated. Thanks in advance.
2026-03-25 15:59:13.1774454353
How to get the following concave surrogate function for this function of a vector?
231 Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in REAL-ANALYSIS
- how is my proof on equinumerous sets
- Finding radius of convergence $\sum _{n=0}^{}(2+(-1)^n)^nz^n$
- Optimization - If the sum of objective functions are similar, will sum of argmax's be similar
- On sufficient condition for pre-compactness "in measure"(i.e. in Young measure space)
- Justify an approximation of $\sum_{n=1}^\infty G_n/\binom{\frac{n}{2}+\frac{1}{2}}{\frac{n}{2}}$, where $G_n$ denotes the Gregory coefficients
- Calculating the radius of convergence for $\sum _{n=1}^{\infty}\frac{\left(\sqrt{ n^2+n}-\sqrt{n^2+1}\right)^n}{n^2}z^n$
- Is this relating to continuous functions conjecture correct?
- What are the functions satisfying $f\left(2\sum_{i=0}^{\infty}\frac{a_i}{3^i}\right)=\sum_{i=0}^{\infty}\frac{a_i}{2^i}$
- Absolutely continuous functions are dense in $L^1$
- A particular exercise on convergence of recursive sequence
Related Questions in OPTIMIZATION
- Optimization - If the sum of objective functions are similar, will sum of argmax's be similar
- optimization with strict inequality of variables
- Gradient of Cost Function To Find Matrix Factorization
- Calculation of distance of a point from a curve
- Find all local maxima and minima of $x^2+y^2$ subject to the constraint $x^2+2y=6$. Does $x^2+y^2$ have a global max/min on the same constraint?
- What does it mean to dualize a constraint in the context of Lagrangian relaxation?
- Modified conjugate gradient method to minimise quadratic functional restricted to positive solutions
- Building the model for a Linear Programming Problem
- Maximize the function
- Transform LMI problem into different SDP form
Related Questions in CONVEX-ANALYSIS
- Proving that: $||x|^{s/2}-|y|^{s/2}|\le 2|x-y|^{s/2}$
- Convex open sets of $\Bbb R^m$: are they MORE than connected by polygonal paths parallel to the axis?
- Show that this function is concave?
- In resticted domain , Applying the Cauchy-Schwarz's inequality
- Area covered by convex polygon centered at vertices of the unit square
- How does positive (semi)definiteness help with showing convexity of quadratic forms?
- Why does one of the following constraints define a convex set while another defines a non-convex set?
- Concave function - proof
- Sufficient condition for strict minimality in infinite-dimensional spaces
- compact convex sets
Related Questions in CONVEX-OPTIMIZATION
- Optimization - If the sum of objective functions are similar, will sum of argmax's be similar
- Least Absolute Deviation (LAD) Line Fitting / Regression
- Check if $\phi$ is convex
- Transform LMI problem into different SDP form
- Can a linear matrix inequality constraint transform to second-order cone constraint(s)?
- Optimality conditions - necessary vs sufficient
- Minimization of a convex quadratic form
- Prove that the objective function of K-means is non convex
- How to solve a linear program without any given data?
- Distance between a point $x \in \mathbb R^2$ and $x_1^2+x_2^2 \le 4$
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
A good reference for this is Kenneth Lange's book, MM Optimization Algorithms, SIAM, 2016.
The basic idea here is that we want a surrogate function $g(x|x_{0})$ with the properties:
$f(x_{0})=g(x_{0} |x_{0})$
$\nabla f(x_{0})= \nabla g(x_{0}|x_{0})$
$g(x|x_{0}) \leq f(x)$ for all $x$
$g(x|x_{0})$ is separable (to make maximizing $g$ easy!)
Proof.
It's easy to see that the candidate function given in the problem statement has $g(x_{0}|x_{0})=f(x_{0})$.
Since
$\nabla f(x)=\frac{-2a}{(a\|x\|^{2}+b)^{2}}x$,
we have
$\nabla f(x_{0})=\frac{-2a}{(a\|x_{0}\|^{2}+b)^{2}}x_{0}$
and it is easy to check that $\nabla g(x_{0}|x_{0})=\nabla f(x_{0})$.
$\nabla^{2} g(x|x_{0}) \preceq \nabla^{2} f(x)$ for all $x$.
It's easy to show that
$\nabla^{2} f(x)=\frac{-2a}{(a\| x \|^{2}+b)^{2}}I + \frac{8a^{2}}{(a\|x \|^{2}+b)^{3}}xx^{T}$
while
$\nabla^{2} g(x|x_{0})=\frac{-2a}{b^{2}}I$.
Thus the eigenvalues of $\nabla^{2}g(x|x_{0})$ are all smaller than the eigenvalues of $\nabla^{2} f(x)$.
We can write a Taylor series with remainder term for $f(x)-g(x|x_{0})$ centered at $x_{0}$ as
$f(x)-g(x|x_{0})= (f(x_{0}-g(x_{0}|x_{0})) + (\nabla f(x_{0}) - \nabla g(x_{0}|x_{0}))^{T}(x-x_{0})+ \frac{1}{2} (x-x_{0})^{T} \left( \nabla^{2}f(\xi)-\nabla^{2}g(\xi |x_{0}) \right) (x-x_{0})$
for some $\xi$ on the line segment between $x$ and $x_{0}$.
$f(x)-g(x|x_{0})=0 + 0 + \frac{1}{2} (x-x_{0})^{T} \left( \nabla^{2}f(\xi)-\nabla^{2}g(\xi |x_{0}) \right) (x-x_{0})$.
Since $\nabla^{2} f(\xi)- \nabla^{2} g(\xi | x_{0})$ is a positive semidefinite matrix,
$\frac{1}{2} (x-x_{0})^{T} \left( \nabla^{2}f(\xi)-\nabla^{2}g(\xi |x_{0}) \right) (x-x_{0}) \geq 0$
and
$f(x)-g(x|x_{0}) \geq 0$ for all $x$.
This general approach to finding a separable concave underestimating surrogate is used frequently in EM and MM methods.
P.S. I've been asked how you could derive this from scratch. The general process would be:
$\frac{p}{2} \| x\|^{2}$.
This ensures that
$\nabla^{2} g(x|x_{0})=pI \preceq \nabla^{2}f(x)$.
$\nabla g(x|x_{0})=px+c$
Solve for $c$ by setting $\nabla g(x_{0}|x_{0})=\nabla f(x_{0})$ to get
$px_{0} + c=\nabla f(x_{0})$
$c=\nabla f(x_{0})-px_{0}$.
$g(x_{0}|x_{0})=\frac{p}{2}\|x_{0}\|^{2}+c^{T}x_{0}+d=f(x_{0})$.
$d=f(x_{0})- \frac{p}{2}\|x_{0}\|^{2} -c^{T}x_{0}$.