I am trying to minimize the function $f(x) = \sqrt[x]{c\times x!}$ with respect to $x$($c>0$ is a constant), for $x>0$. I have tried plotting the function to examine its behavior, and it looks to me that if we consider values of $x>0$, this function should have a global minimum. However, I don't know how to solve this minimization problem analytically, as the derivative is a rather strange looking function.
2026-04-12 16:56:18.1776012978
Minimize a function
407 Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in CALCULUS
- Equality of Mixed Partial Derivatives - Simple proof is Confusing
- How can I prove that $\int_0^{\frac{\pi}{2}}\frac{\ln(1+\cos(\alpha)\cos(x))}{\cos(x)}dx=\frac{1}{2}\left(\frac{\pi^2}{4}-\alpha^2\right)$?
- Proving the differentiability of the following function of two variables
- If $f ◦f$ is differentiable, then $f ◦f ◦f$ is differentiable
- Calculating the radius of convergence for $\sum _{n=1}^{\infty}\frac{\left(\sqrt{ n^2+n}-\sqrt{n^2+1}\right)^n}{n^2}z^n$
- Number of roots of the e
- What are the functions satisfying $f\left(2\sum_{i=0}^{\infty}\frac{a_i}{3^i}\right)=\sum_{i=0}^{\infty}\frac{a_i}{2^i}$
- Why the derivative of $T(\gamma(s))$ is $T$ if this composition is not a linear transformation?
- How to prove $\frac 10 \notin \mathbb R $
- Proving that: $||x|^{s/2}-|y|^{s/2}|\le 2|x-y|^{s/2}$
Related Questions in OPTIMIZATION
- Optimization - If the sum of objective functions are similar, will sum of argmax's be similar
- optimization with strict inequality of variables
- Gradient of Cost Function To Find Matrix Factorization
- Calculation of distance of a point from a curve
- Find all local maxima and minima of $x^2+y^2$ subject to the constraint $x^2+2y=6$. Does $x^2+y^2$ have a global max/min on the same constraint?
- What does it mean to dualize a constraint in the context of Lagrangian relaxation?
- Modified conjugate gradient method to minimise quadratic functional restricted to positive solutions
- Building the model for a Linear Programming Problem
- Maximize the function
- Transform LMI problem into different SDP form
Related Questions in MAXIMA-MINIMA
- optimization with strict inequality of variables
- Minimum value of a complex expression involving cube root of a unity
- Calculation of distance of a point from a curve
- Find all local maxima and minima of $x^2+y^2$ subject to the constraint $x^2+2y=6$. Does $x^2+y^2$ have a global max/min on the same constraint?
- Solving discrete recursion equations with min in the equation
- Trouble finding local extrema of a two variable function
- Why do I need boundedness for a a closed subset of $\mathbb{R}$ to have a maximum?
- Find the extreme points of the function $g(x):=(x^4-2x^2+2)^{1/2}, x∈[-0.5,2]$
- Maximizing triangle area problem
- Find the maximum volume of a cylinder
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
As @themathandlanguagetutor, consider $$f(x)=\Big[c \,\Gamma (x+1)\Big]^{\frac{1}{x}}$$ and assume $c>0$. Using logarithmic differentiation, we have $$\frac{f'(x)}{f(x)}=\frac 1 {x^2}\Big[x\, \psi (x+1)-\log (\Gamma (x+1))-\log (c) \Big]$$ So, we are looking for the zero of function $$g(x)=x\,\psi (x+1)-\log (\Gamma (x+1))-\log (c)$$ $$g'(x)=x \,\psi ^{(1)}(x+1) \quad >0 \quad \quad\forall x >0$$ So, only one solution.
Now, $\big[x\,\psi (x+1)-\log (\Gamma (x+1))\big]$ is very linear as soon as $x > 1$ and a good approximation of it is $$x\,\psi (x+1)-\log (\Gamma (x+1))=x+\frac{1}{2} (1-\log (2 \pi x))-\frac{1}{6 x}+O\left(\frac{1}{x^2}\right)$$ So, keeping the first and second terms, as an approximation, we need to solve $$x+\frac{1}{2} (1-\log (2 \pi x))=\log(c) \implies x=-\frac{1}{2} W_{-1}\left(-\frac{e}{\pi c^2}\right)$$ where $W_{-1}(.)$ is the second branch of Lambert function.
This would be the $x_0$ for Newton method.
Below are some results
$$\left( \begin{array}{ccc} c & x_0 & \text{solution} \\ 2 & 1.20556 & 1.39437 \\ 3 & 1.81582 & 1.93161 \\ 4 & 2.19930 & 2.29176 \\ 5 & 2.48314 & 2.56345 \\ 6 & 2.70899 & 2.78165 \\ 7 & 2.89662 & 2.96393 \\ 8 & 3.05712 & 3.12043 \\ 9 & 3.19732 & 3.25750 \\ 10 & 3.32177 & 3.37942 \\ 20 & 4.12296 & 4.16825 \\ 30 & 4.58111 & 4.62144 \\ 40 & 4.90271 & 4.94016 \\ 50 & 5.15051 & 5.18600\\ 60 & 5.35202 & 5.38606 \\ 70 & 5.52178 & 5.55470 \\ 80 & 5.66842 & 5.70041 \\ 90 & 5.79746 & 5.82868 \\ 100 & 5.91266 & 5.94323 \end{array} \right)$$ To give an idea about the convergence, let us try for $c=10^4$; the iterates would be (with a ridiculous number of figures) $$\left( \begin{array}{cc} n & x_n \\ 0 & 10.819975929577196143 \\ 1 & 10.836092659850328733 \\ 2 & 10.836092115380538490 \\ 3 & 10.836092115380537870 \end{array} \right)$$
So, we have a simple way to locate the minimum. Now, computing $f(x)$ does not make any problem.
Edit
Knowing the approximation $x_0$ we can improve a lot using one single iteration of Newton, Halley or Householder method. The simplest one will be $$x_1=x_0-\frac {g(x_0)}{g'(x_0)}$$ The above table will now be $$\left( \begin{array}{ccc} c & x_1 & \text{solution} \\ 2 & 1.39889 & 1.39437 \\ 3 & 1.93246 & 1.93161 \\ 4 & 2.29214 & 2.29176 \\ 5 & 2.56368 & 2.56345 \\ 6 & 2.78180 & 2.78165 \\ 7 & 2.96405 & 2.96393 \\ 8 & 3.12052 & 3.12043 \\ 9 & 3.25758 & 3.25750 \\ 10 & 3.37948 & 3.37942 \\ 20 & 4.16828 & 4.16825 \\ 30 & 4.62146 & 4.62144 \\ 40 & 4.94018 & 4.94016 \\ 50 & 5.18601 & 5.18600\\ 60 & 5.38607 & 5.38606 \\ 70 & 5.55470 & 5.55470 \\ 80 & 5.70042 & 5.70041 \\ 90 & 5.82869 & 5.82868 \\ 100 & 5.94323 & 5.94323 \end{array} \right)$$
In other words, for practical applications, one single iteration of Halley or Householder methods will give the solution.
For illustration, for the case of $c=10^4$, here are the values ofhe first iterate as a function of the order of the method $$x_1^{(2)}=\color{red}{10.836092}6598503$$ $$x_1^{(3)}=\color{red}{10.83609211}48391$$ $$x_1^{(4)}=\color{red} {10.83609211538}11$$ $$x_1^{(\infty)}=\color{red}{10.83609211538054}$$
It must be noticed that $f(x_0)\times f''(x_0) <0 \quad \forall c$ which, by Darboux theorem, means that we shall face one overshoot of the solution along the path to convergence. At the opposite $f(x_1)\times f''(x_1) >0 \quad \forall c$ which will avoid it.