Suppose $F=\mathbb{R}$. Let $A: V\to V$ (where $V$ is a finite dimensional inner product space over $F$) so that $A=A^*$ ("self-adjoint"), then there exists an orthonormal basis of eigenvectors and eigenvalues are real.
(I come across the above problem when trying to understand how one could represent adjoint operators via their orthonormal base. I cannot find a complete proof in any books (from linear algebra to functional analysis) I hope someone could help in providing a full proof.)
Thanks!
Suppose $V$ is a finite-dimensional inner product space over the real numbers, and suppose $A$ is a symmetric linear operator on $V$. One way to find an eigenvector of $A$ is to minimize $F(v)=(Av,v)$ subject to the constraint that $\|v\|=1$. Such a minimum exists because you can use an orthonormal basis of $V$ and turn this into a minimization problem for a continuous function $F$ over the unit spherical shell in $\mathbb{R}^{n}$, which is compact. A continuous function on a compact set in $\mathbb{R}^n$ has a minimum.
Suppose the minimum of $F$ on $S=\{ v \in V : \|v\|=1 \}$ is $\lambda$, and suppose $F(v_0)=\lambda$ for some unit vector $v_0$. Then you can show that $\lambda$ is an eigenvalue and $(A-\lambda I)v_0=0$. This is because you must have $$ \lambda \le \frac{(Aw,w)}{(w,w)},\;\;\; w\ne 0 \\ 0 \le ((A-\lambda I)w,w),\;\;\; w \in V. $$ That means $[x,y]=((A-\lambda I)x,y)$ has all of the properties of an inner product except that $[x,x]=0$ may occur even if $x \ne 0$. You can check that $[w,w] \ge 0$ because of the above, and it is easy to check that $[x,y]$ is linear in both $x$ and $y$. That's enough in order for the Cauchy-Schwarz inequality to hold: $$ |[x,y]| \le [x,x]^{1/2}[y,y]^{1/2}\\ |((A-\lambda I)x,y)| \le ((A-\lambda I)x,x)^{1/2}((A-\lambda I)y,y)^{1/2} $$ But you know that $[v_0,v_0]=((A-\lambda I)v_0,v_0)=0$ because of how $v_0$ was constructed. Therefore, $$ ((A-\lambda I)v_0,y)=0, \;\;\; y \in V. $$ This is true for all $y\in V$. Taking $y=(A-\lambda I)v_0$ gives $(A-\lambda I)v_0=0$.
Finally, you can use finite induction to get everything. That's because you can now reduce a symmetric operator $A$ on a subspace of one lower dimension. This is because $V' = \{ v : (v,v_0)=0 \}$ is invariant under $A$. To see why, suppose $v \perp v_0$. Then $(Av,v_0) = (v,Av_0)=(v,\lambda v_0)=\lambda(v,v_0)=0$. So $A : V'\rightarrow V'$ and the restriction $A'$ of $A$ to $V'$ is symmetric. Keep repeating the above procedure and eventually you end up with an orthonormal basis of eigenvectors of $V$.