$\newcommand{\norm}[1]{\|#1\|}$ I have the following problem: $$ \min_{w,\theta}\frac{1}{2}\norm{w-w_t}^2+\frac{1}{2}(\theta-\theta_t)^2 \text{ s.t. } w^\top(z(n-\theta)-\hat z(\hat n - \theta)) \ge 1 $$ This problem is at the core of a learning algorithm that attempts to recover a weight vector $w$ and a parameter $\theta$, given some constraints. Here there is just one constraint.
A much simpler version of this problem does not involve $\theta$: $$ \min_w\frac{1}{2}\norm{w-w_t}^2\text{ s.t. }w^\top(z-\hat z)\ge 1. $$ This simpler problem has a closed form that can be retrieved with Lagrange multipliers: basically, let $$ L(w,\lambda)=\frac{1}{2}\norm{w-w_t}^2+\lambda (w^\top(z-\hat z)-1) $$ Then, cancel $\nabla_wL=w-w_t+\lambda(z-\hat z)=0$, then express $\lambda$ as a function of $w$, by cancelling $\nabla_\lambda L$: $$ \lambda=\frac{w_t-(z-\hat z)-1}{\norm{z-\hat z}^2} $$ This allows you to obtain $w$: $$ w=w_t-\frac{w_t^\top(z-\hat z)-1}{\norm{z-\hat z}^2}(z-\hat z) $$
The problem is when I try to do the same thing with the first problem, I have a quite complicated expression for $\lambda$. Something like $a\lambda^3+b\lambda^2+c\lambda+d=0$. Then I am stuck.
Anyone has a suggestion ?
Solve the problem for fixed $\theta$ as you described; then optimise with respect to $\theta$ without constraints.