I wish to find the most likely estimator of the precision matrix (inverse covariance matrix). One option is to maximise the following: $$ f(\Theta) = \frac{N}{2}\log|\Theta|-\sum_i \mathbf{x}_i^T\Theta\mathbf{x}_i $$ (assume that the mean is zero without loss of generality). $\mathbf{x}_i$'s are constant.
I know how to differentiate the above by exploiting the fact that $\mathbf{x}_i^T\Theta\mathbf{x}_i=Tr(\Theta\mathbf{x}_i\mathbf{x}_i^T)$. However, if I pose the question as instead to optimise the cholesky decomposition $L$ where $LL^T=\Theta$, $$ f(L) = {N}\log|L|-\sum_i \mathbf{x}_i^TLL^T\mathbf{x}_i $$ what is $\frac{\partial f(L)}{\partial L}$? It's really the second term that I am struggling with.
$\dfrac{\partial}{\partial L_{jk}} (x^T L L^T x) = 2 x_j (L^T x)_k$