TL;DR: Given $f:\mathbb R \to \mathbb R^n$ smooth enough, what are $df$ and $d^2f$ and how are they computed wrt possibly moving bases $\{e_i(t)\}$? I (believe that I) understand the answers in the "reverse" case where $g:\mathbb R^n \to \mathbb R$ - then $dg$ is the second component of the natural extension of $f$ to a fiber bundle homomorphism $T\mathbb R^n\to T\mathbb R$, that is, $dg:\mathbb R^n \to (R^n)^*$, and $d^2g:\mathbb R^n\to(R^n)^*\wedge (R^n)^*$, or $dg\in A^1(T\mathbb R^n)$ and $d^2g\in A^2(T\mathbb R^n)$. But with my $f$, the $df$ and $d^2f$ just don't "type check" because the "star" gives scalar-valued functions, not vector-valued ones.
Long story: I'm reading a physics book and I just can't get my head around what they're doing. Assume every function here to be $C^\infty$. Using Einstein notation, i.e. every expression $a^i b_i$ is meant to read $\sum_i a^ib_i$.
Setup. In a smooth manifold $M\subseteq \mathbb R^n$ we have a rigid body moving within a time interval $I\subseteq \mathbb R$. The body movement is described by a point $\mathbf R:I\to M$ and a coordinate system (basis) $e_i : I\to \mathbb R^n,i=1,\ldots,n$, thought of as "attached" to the body for each $t\in I$ at the point $\mathbf R(t)$. For brevity I'll omit the dependence on time and write $\mathbf R$ even if $\mathbf R(t)$ is meant (and $e_i$ instead of $e_i(t)$).
Denote the standard basis in $\mathbb R^n$ with $\boldsymbol{\epsilon}_{i}$. Now $\mathbf R=R^i \boldsymbol\epsilon_i$ for some $R^i(t),i=1,\ldots,n$. Since (for each $t\in I$) $\{e_i(t)\}$ forms a basis in $\mathbb R^n$, too, we can express $\boldsymbol{e}_{i}=g_{i}^{j}(\mathbf{R})\boldsymbol{\epsilon}_{j}$ and $\boldsymbol{\epsilon}_{i}=\left(g^{-1}(\mathbf{R})\right)_{i}^{j}\boldsymbol{e}_{j}$ for some invertible matrices $g(\textbf R(t))$.
First differential. Next in the book they take the differential of $\mathbf R$ and do some computations: $$ d\mathbf{R} =d\left(R^{i}\boldsymbol{\epsilon}_{i}\right) =\left(dR^{i}\right)\boldsymbol{\epsilon}_{i} =\left(dR^{i}\right)\left(g^{-1}(\mathbf{R})\right)_{i}^{j}\boldsymbol{e}_{j}$$
I'm uneasy about the second equality $d\left(R^{i}\boldsymbol{\epsilon}_{i}\right) =\left(dR^{i}\right)\boldsymbol{\epsilon}_{i}$ transforming the differential of a vector into a vector multiplied by the differential of a scalar function. Though suspicious, at least symbolically it seems legit because $\boldsymbol{\epsilon}_{i}$ does not depend on the time $t\in I$, so it somehow makes sense to reduce the action of $d$ only to the scalar multiplier $R^i$.
What is the nature of the product between a differential 1-form $dR^i$ and a vector $\boldsymbol \epsilon_i$?
Second differential. Here's where it gets too wild for me to move forward easily. They take the second differential of $\mathbf{R}$ like this: $$0=dd\mathbf{R}=d(\omega^j \boldsymbol{e}_{j})=d\omega^j \boldsymbol{e}_{j}-\omega^j \wedge d\boldsymbol{e}_{j}$$ where the $\omega^j$ are defined s.t. $d\mathbf{R}=\omega^j \boldsymbol{e}_{j}$. We have now a product of a two-form $d\omega^j$ with a vector $\boldsymbol{e}_{j}$, plus a wedge product of a one-form $\omega^j$ with a "vector-form" $d\boldsymbol{e}_{j}$, and both are supposed to be of the same type.
$\omega^j \wedge d\boldsymbol{e}_{j}$ looks like complete non-sense to me, because a wedge product can take only same-type objects. There is $V\wedge V$ but never $V\wedge W$ for $V\neq W$. And yet, here $\omega^j$ is a "scalar" one-form while $d\boldsymbol{e}_{j}$ is a "vector" one-form, and they're wedged together. And I'm not even sure if I have to write the vector in the second position, it seems if I write $-d\boldsymbol{e}_{j} \wedge \omega^j$ it'd be ok, too.
I suspected something like a "vector-valued form" should exist, and indeed there's a wikipedia article on it. I read through it a few times, but I got too confused with the abstract construction done in two parts - defining a "bundle-valued" form and then taking a tensor product of bundles. Neither could I really understand what are the rules I can use in my more "practical" problem.
My questions are:
- what is the "type signature" of $\mathbf R$ as a (vector-valued) diffferential (0-)form, allowing us take its exterior derivatives $d\mathbf R$ and $d^2\mathbf R$?
- what does it mean to multiply "scalar" and "vector" forms?
- how is such a mixed-type product computed? Why does $d(fg)=df\wedge g-f\wedge dg$ still apply, if $f$ and $g$ are differently typed?
I think you mostly figured out the details, and yes the tensor product allows you to work with things neatly. I just want to emphasize that your questions can all be answered if you have general enough definitions. In what follows, everything is real (vector spaces, vector bundles) and smooth (maps, manifolds, bundles).
1. Products of Vectors in General.
So to go from vector spaces to vector bundles, you simply do things fiberwise (or if you want to think of it the other way, to go from vector bundles to vector spaces, simply take $M$ to be a singleton point).
Here are some common examples:
Doing this fiberwise, you can get relevant vector-bundle examples.
2. Vector-Valued Alternating Forms, and General Wedge Products.
Here, you can check out Ivo Terek’s answer for some details.
Notice how this is really not a complicated thing at all, because the usual definition takes $E=\Bbb{R}$. Also, I should remark that there is no differential geometry here, purely abstract algebra; so a $k$-form here is not a $k$-form “field”. Hopefully context will inform you of the intended meaning.
Note that by the universal property for the exterior powers, this amounts to the same thing as a linear map $\omega:\bigwedge^k(V)\to E$ (strictly speaking I should write something like $\tilde{\omega}$ to distinguish from the $\omega$ above, but I’ll run out of letters at this rate), or more simply said, it is an element of $\text{Hom}\left(\bigwedge^k(V),E\right)$. By the usual tensor-product and Hom isomorphisms, this space is isomorphic to $\left(\bigwedge^k(V)\right)^*\otimes E\cong \bigwedge^k(V^*)\otimes E$. Of course, algebraically, it doesn’t matter which concrete realization we work with, but I personally just think of them as alternating multilinear maps with values in $E$, or as a linear mapping $\bigwedge^k(V)\to E$.
Now, we come to a definition of a wedge-product for vector-valued forms:
One of the comments you raised is:
$\omega^j\wedge de_j$ is not nonsense, but you’re entirely correct with the second half. But, if you look at the definition, the domains of these maps are all the same ($\bigwedge^k(V)$ for some $k$), it is only their target space which changes ($E$ or $F$ or $G$). So, for exterior algebra, we can freely mess around with the target as long as we have a product $\beta$. But, what I can’t do is define a wedge product of an element $\text{Hom}\left(\bigwedge^k(V),E\right)$ with an element in $\text{Hom}\left(\bigwedge^klW),F\right)$ if $V\neq W$.
Now, here are some properties easily proved from the definitions (or by opening your favourite book and adapting the proofs (actually just copying line by line and modifying notation appropriately)):
One example of such a collection of compatible products is scalar multiplication in the field of course. A non-trivial example is when $E_1$ is a space of linear maps, $E_2$ is another space of linear maps, and $E_3$ is a space of vectors such that composition and evaluation are defined; then obviously $T(S(x))=(T\circ S)(x)$. At the level of forms, this gives \begin{align} \tau\wedge_{\text{ev}}(\sigma\wedge_{\text{ev}}\xi)=(\tau\wedge_{\circ}\sigma)\wedge_{\text{ev}}\xi. \end{align} Here, $\text{ev}$ appears three times, but it’s not necessarily the same evaluation map of course. One important special case is when you consider $\text{ev}:\text{End}(E)\times E\to E$ and $\circ:\text{End}(E)\times\text{End}(E)\to\text{End}(E)$. Then, the same formula holds of course, and the three $\text{ev}$ maps are the same.
So now, with the bilinearity of $\wedge_{\beta}$, and the theorems above of anti-commutativity, associativity, and distributivity under pullback (extremely important), you can now freely manipulate vector-valued forms as you would with the scalar-valued case. You just need to know which bilinear product is intended at each stage.
3. Vector-bundle valued differential forms
This is a very simple thing: for each point $x$ in the base, we require $\omega$ to restrict to a linear map $\bigwedge^k(T_xM)\to E_x$. So, we’re simply applying the definition above fiber-by-fiber. There are two noteworthy special cases:
One final definition I’ll add is that of the pullback because there’s one subtlety:
The reason pullbacks here are slightly more nuanced is because on the target, we don’t just land in a single vector space.
With these definitions in mind, you can generalize all the theorems mentioned in section 2 to the case of differential forms taking values in a vector bundle (and hopefully you can extend the definition of pullback to this case as well; just do it fiber-by-fiber).
4. Exterior derivatives for vector-space valued differential forms.
When talking about differentiation, we have roughly speaking, three situations:
Here are the properties of the such-defined $d$:
So, this all sounds just as in the scalar case, and you’re right. For the product rule, you just put $\wedge_{\beta}$ and you’re good to go, and this should answer your questions about $(dR^i)\,\mathbf{\epsilon}_i$ and also $\omega^j\wedge d\mathbf{e}_j$.
For example, you asked about the meaning of $(d R^i)\,\mathbf{\epsilon}_i$ in your question. Well, in this case, notice that $R^i$ is a smooth function, meaning $V=\Bbb{R}$ above. Next, $\epsilon_i$ is an element of $\Bbb{R}^n$, so $W=\Bbb{R}^n$. The mapping $\beta$ is thus simply the scalar multiplication of a number with a vector in $\Bbb{R}^n$. To be fully explicit, let us completely evaluate. Given a time $t\in\Bbb{R}$ and a “tangent vector at $t$”, $\xi$, we can get $[(dR^i)\,\mathbf{\epsilon_i}]_t(\xi)= (dR^i)_t[\xi]\cdot \mathbf{\epsilon}_i= (R^i)’(t)\cdot \xi\cdot\mathbf{\epsilon}_i$.
This may be confusing because your $R^i$ is a function $I\to\Bbb{R}$ to it may be too specialized to understand the definitions. If you have a usual $k$-form $\omega$ on a manifold $M$, and you have a vector $v\in V$, then the “product” $\omega\cdot v$ can be thought of as the wedge product of the usual scalar form $\omega$ with the constant map $f:M\to V$, $f(x)=v$ (a $0$-form) and you wedge them with respect to scalar multiplication $\Bbb{R}\times V\to V$. The product nature is more apparent once we eliminate the form-nature: let us evaluate $\omega$ fully: take a point $x\in M$ and tangent vectors $\xi_1,\dots,\xi_1\in T_xM$. Then we get a number $\omega(\xi_1,\dots,\xi_k)$. This number can be multiplied by the vector $v$. So, \begin{align} (\omega\cdot v)(\xi_1,\dots, \xi_k)&=\omega(\xi_1\dots,\xi_k)\cdot v. \end{align} So, the thing to keep in mind is that “the product takes place in the target, not the domain”.
5. Exterior covariant derivative.
Let $(E,\pi,M)$ be a vector bundle, and suppose $E$ is equipped with a connection $\nabla$. Then, given a section $\psi:M\to E$ (which is also an $E$-valued $0$-form on $M$), you can consider not just the tangent map $T\psi:TM\to TE$, but also the covariant derivative $\nabla\psi:TM\to E$ (which is essentially a “projection” of $T\psi$, but for general vector bundles there’s no canonical way to define this projection, hence we need the connection). Its value on a vector $h_x\in T_xM$ gives you a vector $\nabla_{h_x}\psi\in E_x$. Notice that $\nabla\psi$ is precisely an $E$-valued $1$-form on $M$.
Now, rather than writing $\nabla\psi$, it is sometimes useful to write it as $d_{\nabla}\psi$. This reminds us that $\psi$ is an $E$-valued $0$-form on $M$ and $d_{\nabla}\psi$ is an $E$-valued $1$-form on $M$. One can extend the definition of $d_{\nabla}$ so it acts on all forms, to give a sequence of linear maps $d_{\nabla}:\Omega^k(M;E)\to\Omega^{k+1}(M;E)$. Here are the properties of $d_{\nabla}$:
In other words, if you omit the superscripts $E,F,G$, then the exterior covariant derivative $d_{\nabla}$ behaves as expected. Note that the compatibility condition is satisfied for many “products” of interest, for example for metric-compatible connections in Riemannian/Lorentzian geometry. Or for example when $\beta$ is evaluation $\text{Hom}(E,F)\oplus E\to F$.
6. Fully general calculations using all this machinery
Let $f:N\to M$ be a smooth map between smooth manifolds, and $(E,\pi,M)$ a vector bundle equipped with a connection $\nabla$, and consider the pullback bundle $(f^*E,\pi,N)$ with the induced pullback connection, which for convenience I shall continue to denote as $\nabla$.
So, my map $f$ is meant to generalize your $\mathbf{R}:I\to M$. Notice I’m not making any assumptions about my $M$. I don’t care what ambient space it lives in; I simply care that I have a vector bundle with a connection. For example, you can take $E=M\times V$ for a fixed vector space $V$, such as the ambient $\Bbb{R}^n$ in your example, which would then simplify everything below since I no longer need connections. Or you could take $E=TM$, or something fancier if you wish.
Now, suppose I have a local frame $\{e_1,\dots, e_p\}$ for the pullback bundle $f^*E$ over some open set $U\subset N$. Since all my considerations below are local with respect to $N$, I may as well assume $N$ is small enough, and that $U=N$. This essentially means for each point $z\in N$, I have a basis $\{e_1(z),\dots, e_p(z)\}$ for the vector space $E_{f(z)}$. Intuitively, my mapping $f$ maps the point $z\in N$ to $f(z)\in M$ and I now attach a basis at this point for the fiber $E_{f(z)}$. Now, I can introduce the dual frame $\{\sigma^1,\dots,\sigma^p\}$ to the $e$’s. Essentially, this means for each point $z\in N$, $\{\sigma^1(z),\dots, \sigma^p(z)\}$ is the dual basis for $E_{f(z)}^*$ with respect to the basis $\{e_1(z),\dots, e_p(z)\}$ of $E_{f(z)}$.
Ok, so now each $e_j$ is a section of the bundle $f^*E$ over $N$, i.e an $f^*E$-valued $0$-form on $N$. Thus, I can take its covariant exterior derivative $d_{\nabla}(e_j)$ to get an $f^*E$-valued $1$-form on $N$ (again, the $\nabla$ here is really $f^*\nabla$). I can now apply $\sigma^i$ to this to get an honest to god 1-form on $N$, $\omega^i_{\,j}:=\sigma^i(d_{\nabla}e_j)$. If you want to fit this into the above framework, I’m taking $\beta$ to be the duality pairing (which is a form of evaluation) between the $(f^*E)^*$-valued $0$-form $\sigma^i$ and the $f^*E$-valued $1$-form $d_{\nabla}(e_j)$. Anyway, the point of these scalar forms (also called the connection forms) is that I can re-express the exterior covariant derivative in terms of the $e$’s again: \begin{align} d_{\nabla}(e_j)&=\omega^i_{\,j}\cdot e_i, \end{align} where here the $\cdot$ is the product associated to scalar multiplication (note I don’t write $\wedge_{\cdot}$ here because the $e_i$ are ($E$-valued) $0$-forms, but I wouldn’t be wrong if I wrote $d_{\nabla}(e_j)=\omega^i_{\,j}\wedge_{\cdot}e_i$).
If we keep going, then we can apply $d_{\nabla}$ to this equation to get \begin{align} d_{\nabla}^2(e_j)&=d(\omega^i_{\,j})\cdot e_i+ (-1)^1\omega^i_{\,j}\wedge (d_{\nabla}e_j). \end{align}
I leave it to you to carefully figure out what the $E,F,G$ are and what the $\nabla^E,\nabla^F,\nabla^G$ are, in the notation of my theorem above about product rule for vector-bundle valued forms, and why this justifies the equality above. Now, we can simplify further; on the LHS the second exterior covariant derivative is no longer $0$, but actually depends on the curvature. So, if I let $R$ be the curvature of $\nabla$ on $E$ (which is an $\text{End}(E)$-valued $2$-form on $M$), and I let $R_f$ be the curvature of $f^*\nabla$ on $f^*E$, one can show that $R_f=f^*R$ (pull-back of an endomorphism-valued 2-form) and that the LHS of the equation above is $R_f\cdot_{\text{ev}} e_j$, where this $\cdot_{\text{ev}}$ is the product associated to evaluation of an endomorphism on a vector. So, \begin{align} R_f\cdot_{\text{ev}} e_j&= d(\omega^i_{\,j})\cdot e_i -\omega^i_{\,j}\wedge (d_{\nabla}e_j), \end{align} and this is an equality of $f^*E$-valued $2$-forms on $N$.
This is all essentially the beginning of Cartan’s moving frames method, and if you specialize to the tangent bundle and consider Levi-Civita connection, you easily arrive at his two structure equations.
Edit: A rapid way of introducing the algebraic aspects.
As a matter of principle of illustrating the power of tensor products, let’s now give the rapid definitions. Let $V,W,X$ and $E,F,G$ be vector spaces, and suppose we have two bilinear maps (i.e products) $\alpha:V\times W\to X$ and $\beta:E\times F\to G$. By the universal property for tensor products, you now get linear maps $\tilde{\alpha}:V\otimes W\to X$ and $\tilde{\beta}:E\otimes F\to G$. Now, you can take the tensor product of these linear maps to give you a linear map $\tilde{\alpha}\otimes\tilde{\beta}:(V\otimes W)\otimes (E\otimes F)\to X\otimes G$. Now, by permuting the factors, this linear map is essentially the same as the linear map $(V\otimes E)\otimes (W\otimes F)\to X\otimes G$. And of course such a linear map comes from a bilinear map $(V\otimes E)\times (W\otimes F)\to X\otimes G$. Explicitly, the action on pure tensors is \begin{align} \left((v,e), (w,f)\right)\mapsto \alpha(v,w)\otimes \beta(e,f)\equiv (v\cdot_{\alpha}w)\otimes(e\cdot_{\beta}f). \end{align} In other words, we defined a product between $V\otimes E$ and $W\otimes F$ using $\alpha$ and $\beta$. We can think of $V\otimes E$ as being the space of “$E$-valued vectors in $V$” or “vectors in $V$ of type $E$”.
This is a very general situation, and by taking $V,W,X,E,F,G,\alpha,\beta$ carefully, you can recover so many examples. The most relevant for now is: