I am curious about computing the root system (as linear functionals) using the minimum prerequisite knowledge and straightforward computation. For example, for Type A the eigenvalues of $ad_{H}$ can be directly computed.
In Hall's Lie Groups, Lie Algebras, and Representations: An Elementary Introduction it is computed by complexification.
In Rossman's Linear Groups computing Type B-D is skipped.
From my understanding, I felt like the process is (i) Fix an appropriate Cartan subalgebra (ii) Finding out the roots(as eigenvals) of $ad_{H}$. (iii) Find out the corresponding root vectors and done. (*)
Unfortunately, I haven't found many working examples and most resources just tell you what the roots and root vectors are directly. So I'm not sure whether (*) will work. After a few trials, I realized many problems might occur in this process, for one, the Cartan subalgebra is not unique and there are many conjugates of them.
Can somebody enlighten me?
I have been feeling a bit rusty with Lie algebra basics lately, so I decided to do $\mathfrak{so}_5$ as a refresher. Also as an excuse to review some material from Humphreys (R.I.P. — with my deepest eternal respect).
To avoid the trap of using shortcuts like assuming that we get a Cartan subalgebra by including the diagonal matrices, I picked the $SO_5(\Bbb{R})$-route with the preserved quadratic form coming from the Euclidean norm. In other words, I will declare $$ L:=\mathfrak{so}_5=\{A\in M_{5\times5}(\Bbb{C})\mid A^T=-A\}. $$ You see that I did complexify the Lie algebra. Because complex eigenvalues appear, this is prudent.
For all pairs of indices, $1\le i,j\le 5$, I denote by $E_{ij}$ the matrix with a single non-zero entry $1$ at position $(i,j)$. Then a basis for $L$ consists of the $10$ matrices $$ S_{ij}:=E_{ij}-E_{ji}, 1\le i<j\le5. $$ To simplify the formulas I also adopt the convention $S_{ij}:=-S_{ji}$ whenever $j<i$. It is then a simple matter to calculate the commutator formula $$ [S_{ij},S_{k\ell}]=\delta_{jk}S_{i\ell}+\delta_{i\ell}S_{jk}+\delta_{ki}S_{\ell j}+\delta_{\ell j} S_{ki}. $$ But all we really need is the realization that if $\{i,j\}\cap\{k,\ell\}=\emptyset$ then $S_{ij}$ and $S_{k\ell}$ obviously commute, and if the index pairs intersect in a singleton set, then we have a copy of $\mathfrak{so}_3$ with the familiar commutator relations $$ [S_{12},S_{23}]=S_{13},\quad [S_{23},S_{31}]=S_{21},\quad [S_{31},S_{12}]=S_{32}, $$ where we can put any distinct indices $i,j,k$ in place of $1,2,3$.
Next I want to prove the claim that all the elements $S_{ij}, 1\le i<j\le5$ are $\mathrm{ad}$-semisimple. Or, the linear transformations $\mathrm{ad}(S_{ij}):L\to L, x\mapsto [S_{ij},x]$ are diagonalizable. Let's look at $\mathrm{ad}(S_{12})$, the others are gotten with the appropriate index substitutions. While at it, I will write down the eigenspaces of $\mathrm{ad}(S_{12})$ as we will be needing them shortly. I will denote by $V_\lambda$ the eigenspace belonging to the eigenvalue $\lambda$:
The dimensions of the eigenspaces add up to $10=\dim L$, so semisimplicity of $\mathrm{ad}(S_{12})$ follows.
Similarly, let me denote by $W_\mu$ the eigenspaces of $\mathrm{ad}(S_{34})$. These are (do the obvious index substitutions): $$\begin{aligned} W_0&=\langle S_{34},S_{12},S_{15},S_{25}\rangle,\\ W_i&=\langle S_{35}+iS_{45},S_{31}+iS_{41},S_{32}+iS_{42}\rangle,\\ W_i&=\langle S_{35}-iS_{45},S_{31}-iS_{41},S_{32}-iS_{42}\rangle. \end{aligned} $$
As $S_{12}$ and $S_{34}$ are both (ad-)semisimple, and commute, they span a toral subalgebra $$H=\langle S_{12},S_{34}\rangle.$$ It is not immediately obvious that this is a maximal toral subalgebra, but that claim does follow from the fact that any toral subalgebra is abelian (Humphreys, p.35), and the observation that $H=V_0\cap W_0$. After all, $V_0\cap W_0$ is the subspace of $L$ consisting of the elements that commute with both $S_{12}$ and $S_{34}$.
The root spaces (with respect to this choice of $H$) are then the common eigenspaces of $\mathrm{ad}(S_{12})$ and $\mathrm{ad}(S_{34})$ such that at least one of the eigenvalues is non-zero.
We very easily spot that in this case all the intersections $V_\lambda\cap W_\mu$, $(\lambda,\mu)\neq(0,0)$, are actually non-trivial and (as expected) $1$-dimensional: $$ \begin{aligned} V_0\cap W_i&=\langle S_{35}+iS_{45}\rangle,\\ V_0\cap W_{-i}&=\langle S_{35}-iS_{45}\rangle,\\ V_i\cap W_0&=\langle S_{15}+iS_{25}\rangle,\\ V_{-i}\cap W_0&=\langle S_{15}-iS_{25}\rangle,\\ V_i\cap W_i&=\langle S_{13}+iS_{14}+iS_{23}-S_{24}\rangle,\\ V_i\cap W_{-i}&=\langle S_{13}-iS_{14}+iS_{23}-S_{24}\rangle,\\ V_{-i}\cap W_i&=\langle S_{13}+iS_{14}-iS_{23}+S_{24}\rangle,\\ V_{-i}\cap W_{-i}&=\langle S_{13}-iS_{14}-iS_{23}+S_{24}\rangle.\\ \end{aligned} $$ So if we identify $\lambda\in H^*$ with the vector $(\lambda(S_{12}),\lambda(S_{34}))$, the roots are $(\pm i,0)$, $(0,\pm i)$, $(\pm i,\pm i)$ (all four sign combinations in the last). Denoting $$ \alpha=(i,-i)\quad\text{and}\quad\beta=(0,i) $$ we can rewrite the roots as $\alpha,\beta,\alpha+\beta,\alpha+2\beta$ and their negatives, so we have a copy of the root system $B_2$.