-
Notifications
You must be signed in to change notification settings - Fork 7
/
chapter6b.tex
3114 lines (2707 loc) · 173 KB
/
chapter6b.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
\begin{comment}
\section{Linear transformations and tensors}
\begin{Remark}
The set of all invertible linear transformations of a vector space $V$ into itself is called the general linear group $GL(V)$ for that space. $ GL(n,\mathbb{R})$ is the space of invertible $n\times n$ matrices, which acts on $R^n$ as a transformation group $GL(\mathbb{R}^n)$ by matrix multiplication. Tensor transformation laws extend this action to the tensor spaces above $\mathbb{R}^n$, which we will understand only after we know what a tensor is. Both of the above 1-dimensional matrix groups (one free group parameter) are subgroups of $GL(2,\mathbb{R})$. In fact both are subgroups of the special linear group $SL(2,\mathbb{R})$ which consists of all unit determinant invertible matrices.
\end{Remark}
Suppose $A : V \rightarrow V$ is a linear transformation of $V$ into itself, i.e., a $V$-valued linear function on $V$, or equivalently a linear function on $V$ with values in V.
For each $i$, the result $A(e_i)$ is a vector with components defined by $A^j{}_{i} \equiv \omega^j (A(e_i))$
(note natural index positions up/down): $A(e_i)\equiv A^j{}_{i}e_j$.
By linearity
$$
A(v)
= A(v^i e_i)
= v^i A(e_i)
= v^i (A^j{}_{i} e_j)
= (A^j{}_{i} v^i) e_j
\qquad \text{or} \qquad
[A(v)]^j = A^j{}_i v^i \,.
$$
The $j$-th component of the image vector is the $j$-th entry of the matrix product of the matrix $\underline {A}$ $\equiv (A^j{}_{i})$ (the row index $j$ is on the left, the column index $i$ is on the right) with the column vector $\underline {v} \equiv (v^i)$. Here the underlined symbol $\underline{A}$ distinguishes the matrix of the linear transformation from the transformation $A$ itself.
This matrix $\underline {A}=(\omega^j(A(e_i)))$ is referred to as the ``matrix of $A$ with respect to the basis $\{e_i\}$." Obviously if you change the basis, the matrix will change. We'll get to that later.
Even if we are not working with ${\mathbb R}^n$, any choice of basis $\{e_i\}$ of $V$
establishes an isomorphism with ${\mathbb R}^n$, namely the $n$-tuple of components of a vector with respect to this basis is a point in ${\mathbb R}^n$---this essentially identifies the basis $\{e_i\}$ of $V$ with the standard basis of ${\mathbb R}^n$. Recall the natural correspondence of quadratic polynomials with $R^3$ explored in the earlier Exercise \ref{exercise:quadpolys}.
Expressing everything associated with an abstract vector space in terms of components with respect to a given basis leads us to matrix notation. Vectors in component form become column matrices, covectors become row matrices, and a linear transformation becomes a square matrix acting by matrix multiplication on the left, while natural evaluation of a covector on a vector is the matrix multiplication of the corresponding row (left) and column (right) matrices
$$
\underline {v} =
\left(
\begin{array}{cc}
v^1 \\
\vdots \\
v^n \\
\end{array}
\right)\,,\
\underline {f}^T = (f_1 \cdot \cdot \cdot f_n)\,,\
\underline {A} = (A^i{}_{j})\,,\
\quad \rightarrow\quad
\left(
\begin{array}{cc}
[A(v)]^1 \\
\vdots \\ \relax
[A(v)]^n \\
\end{array}
\right)
= \underline {A}\, \underline {v} \,,\
f(v)=\underline {f}^T\, \underline {v}\,.
$$
Since $A(v)$ is another vector we can evaluate it on the covector $f$ to get a number which has the triple matrix product representation
$$
f(A(v))= \underline{f}^T\, \underline{A} \, \underline{v}
\,. \qquad\hbox{(row $\times$ square matrix $\times$ column = scalar)}
$$
For every linear transformation $A$, this enables us to define an associated bi-linear real-valued function $\mathbb{A}$ of a pair of arguments consisting of a covector and a vector. Bi-linear simply means linear in each of two arguments. This bi-linear function is
\begin{eqnarray}
\mathbb{A}(f,v) \equiv f(A(v))
&=& (f_i\omega^i) (A^j{}_{k} v^k e_j)
= f_i A^j{}_{k}v^k\omega^i(e_j)
\nonumber\\
&=& f_i A^j{}_{k}v^k \delta^i{}_j
= f_i A^i{}_{k}v^k \,,\
\nonumber
\end{eqnarray}
noting that $A(v)$ is a vector and $f(A(v))$ is a scalar (real number).
For fixed $f$, $\mathbb{A}(f,v)$ is a real-valued linear function of $v$, namely the covector with components $f_i A^i{}_{k}$ (one free down index).
For fixed $v$, it is a real-valued linear function of $f$, namely evaluation on the vector with components $A^i{}_{k} v^k$ (one free up index). This reinterprets the linear transformation $A$ as a bilinear function $\mathbb{A}$ of a covector (first argument) and a vector (second argument), i.e., a ``tensor." Note the component relation
$A^i{}_j=\mathbb{A}(\omega^i,e_j)$. We will notationally identify $A$ and $\mathbb{A}$ once we are more familiar with these matters. Sometimes one writes the linear transformation as $u\to A(u)=\mathbb{A}(\ ,u)=C_{\ds u} \mathcal{A}$, namely as the tensor with only one of its two arguments evaluated, or sometimes as the ``contraction" of the tensor $\mathbb{A}$ with the vector $u$ to indicate its natural evaluation on that argument alone.
In general a ($^p_q$)-tensor over $V$ is simply a real-valued multilinear function of $p$ covector arguments (listed first) and $q$ vector arguments (listed last):
$$
T(\underbrace{f,g,\cdots}_p,\underbrace{v,u,\cdots}_q) \in \mathbb{R} \,.
$$
Listing all the covector arguments first and the vector arguments last is just an arbitrary choice, and later on it will be convenient to allow any ordering.
By definition then, a covector is a ($^0_1$)-tensor over $V$ (1 vector argument, no covector arguments) while a vector is a ($^1_0$)-tensor over $V$ (1 covector argument, no vector argument) recalling that
$v(f) \equiv f(v)$ (the value of a vector on a covector is the value of the covector on the vector).
Thus a linear transformation $A$ has (naturally) a ($^1_1$)-tensor $\mathbb{A}$ over $V$ associated with it. Any time we have a space of linear functions over a vector space, it has a natural linear structure by defining linear combinations of functions through linear combination of values, i.e., is itself a vector space and we can look for a basis. In this case the space of bilinear real-valued functions on the Cartesian product vector space of pairs $(f,v)$ of covectors and vectors is itself a vector space and in the same way that a basis of $V$ determined a basis of the dual space $V^*$, they both together determine a basis of this latter vector space.
Let $V \otimes V^*$ denote this space of ($^1_1$)-tensors over $V$. The symbol $\otimes$ is called the tensor product, explained below. The zero element of this vector space is a multilinear function
$$
\underset{\rm zero\ tensor}{0(f,v)}
= \underset{\rm zero\ number}{0}
\longleftrightarrow
\underset{\rm zero\ matrix}{0^i{}_j}
= \underset{\rm zero\ linear\ transformation}{\omega^i(0(e_j))} = 0 \,,
$$
whose square matrix of components is the zero matrix (note $0(e_j) = 0$ is the zero vector).
Another special element in this space is the evaluation tensor associated with the identity transformation $I\!d(v)=v$
$$
EV\!\!AL(f,v) = f(v) = f_i\delta^i{}_j v^j
\longleftrightarrow
(EV\!\!AL)^i{}_j =\omega(I\!d(e_j)) =\omega^i(e_j) = \delta^i{}_j
$$
whose square matrix of components is the unit matrix $\underline{I}$\,, the index symbol for which has been called the Knonecker delta. $EV\!\!AL$ is sometimes called the unit tensor, and the associated linear transformation of the vector space is just the identity transformation which sends each vector to itself.
To come up with a basis of $V\otimes V^*$ we need a simple definition. Given a covector and a vector we produce a $(^1_1)$-tensor by defining
$$
(v\otimes f)(g,u) \equiv g(v)f(u) = (g_i v^i) (f_j u^j) = g_i(v^i f_j)u^j \,.
$$
Thus $(v^i f_j)$ is the matrix of components of $v\otimes f$, and is
also the result of evaluating this tensor on the basis vectors and dual basis covectors
$$
(v\otimes f)^i{}_j
=(v\otimes f)(\omega^i,e_j) = \omega^i(v) f(e_j) = v^i f_j \,.
$$
The symbol $\otimes$ is called the tensor product and only serves to hold $v$ and $f$ apart until they acquire arguments to be evaluated on. It simply creates a function taking 2 arguments from two functions taking single arguments.
The component expression shows that $v\otimes f$ is clearly bilinear in its arguments $g$ and $u$, so it is a ($^1_1$)-tensor.
In terms of the corresponding matrix notation, given a column matrix $\ul{u}=\langle u^1,\ldots,u^n\rangle$ and a row matrix $\ul{f}^T=\langle f_1|\ldots|f_n\rangle$, then the tensor product corresponds exactly to the other matrix product (column times row instead of row times column)
$$
(u^i f_j)
=
\underbrace{\underbrace{\ul{u}\strut}_{\ds n\times 1}\, \underbrace{\ul{f}^T}_{\ds 1\times n}}_{\ds n\times n}
\quad\text{in contrast with}\quad
\underbrace{\underbrace{\ul{f}^T}_{\ds 1\times n}\, \underbrace{\strut\ul{u}}_{\ds n\times 1}}_{\ds 1\times 1} =f(u)=f_i u^i\,.
$$
Thus the tensor product of a vector and a covector is just an abstract way of representing the multiplication of a column vector on the left by a row vector on the right to form a square matrix, a two-index object created out of two one-index objects.
\begin{Example}\label{example:}\textbf{matrix product and linear function evaluation}
A concrete example can help. The matrix product on the left below is the usual order of a row on the left and a column on the right, resulting in a scalar. The rows and columns of the matrix product on the right below of a column on the left multiplying a row on the right have only 1 entry each respectively so the row-column products are simply products of those two entries.
$$
\ul{f}^T \, \ul{u} =\begin{pmatrix} f_1 & f_2 \end{pmatrix} \begin{pmatrix} u^1 \\ u^2 \end{pmatrix} = f_1 u^1+f_2 u^2 \,,\quad
\ul{u} \, \ul{f}^T = \begin{pmatrix} u^1 \\ u^2 \end{pmatrix} \begin{pmatrix} f_1 & f_2 \end{pmatrix}
= \begin{pmatrix} f_1 u^1 & f_2 u^1 \\ f_1 u^2 & f_2 u^2 \end{pmatrix}
$$
The latter matrix product is the matrix of components of the tensor product $u\otimes f$ of the vector $u$ with the 1-form $f$. Its matrix product with a component vector corresponds to a linear transformation of the vector space.
\end{Example}
We can use the tensor product $\otimes$ to create a basis for $V\otimes V^*$ from a basis $\{e_i\}$ and its dual basis $\{\omega^i\}$, namely the set $\{e_j \otimes \omega^i\}$ of
$n^2=n \times n$ such tensors.
By definition
$$
(e_j\otimes \omega^i)(g,u)
= g(e_j)\omega^i(u)
= g_j u^i = u^i g_j \,.
$$
We can use this to show the two conditions that they form a basis are satisfied:
\begin{enumerate}
\item
{\bf spanning set}:
$$
\mathbb{A}(f,v)= \cdots
= f_j A^j{}_{k} v^k
= A^j{}_{k} v^k f_j
=(A^j{}_{k}\, e_j\otimes \omega^k)(f,v)
$$
since $v^k f_j = (e_j\otimes \omega^k)(f,v)$,
so $\mathbb{A}=A^j{}_{k}\, e_j\otimes \omega^k$ holds since the two bi-linear functions have the same values on all pairs of arguments. The components of the tensor $\mathbb{A}$ with respect to this basis are just the components of the linear transformation $A$ with respect to $\{e_i\}$ introduced above :
$A^j{}_{k} = \omega^j(A(e_k))$.
\item
{\bf linear independence}: if $A^j{}_{k}\, e_j\otimes \omega^k = 0$ (zero tensor) then evaluating both sides on the argument pair $(\omega^m,e_n)$ leads to
\begin{eqnarray}
(A^j{}_{k}\, e_j\otimes \omega^k)(\omega^m,e_n) &=& 0(\omega^m,e_n) = 0
\nonumber\\
&=& A^j{}_{k}\, \omega^m(e_j)\omega^k(e_n)
= A^j{}_{k}\, \delta^m{}_j \delta^k{}_n
\nonumber\\
&=& A^m{}_n \,,
\end{eqnarray}
so since this is true for all possible values of $(m,n)$, all the coefficients must be zero, proving linear independence.
\end{enumerate}
Thus $V \otimes V^*$ is the space of linear combinations of tensor products of vectors with covectors, explaining the notation.
\begin{Example}\label{example:glnRbasis}\textbf{basis of the vector space of $m\times n$ matrices}
We have no notation for the natural basis of the vector space $gl(n,\mathbb{R})$ of $n\times n$ matrices, namely the standard basis of the corresponding $\mathbb{R}^{n^2}$ we get by listing the entries of the matrix row by row as a single 1-dimensional array.
Let $\ul{e}^j{}_i$ be the matrix whose only nonzero entry is a 1 in the $i$th row and $j$th column. Then $\ul{A} = A^i{}_j \ul{e}^j{}_i$ is how we represent the matrix in terms of its entries. The ordering of the indices on $\ul{e}^j{}_i$
allows us to think of this product as having adjacent indices (the $j$'s) being summed over and taking the trace of the result (the $i$s), which are natural matrix kinds of index operations. (The equally acceptable alternative notation would be instead $\ul{A} = A^i{}_j \ul{e}_i{}^j$, but for some reason the first index ordering pleases me more for the interpretational reason I stated.) Then to the matrix $\ul{A}$ corresponds a tensor $\mathbb{A} = A^i{}_j {e}_i\otimes \omega^j$, whose components with respect to this basis are just the corresponding entries of the matrix, so really the basis $\{e_i\otimes \omega^j\}$ of $\mathbb{R}^n \otimes \mathbb{R}^{n*}$ induced by the standard basis $\{e_i\}$ of $\mathbb{R}^n$ corresponds exactly to the obvious basis $\{\ul{e}^j{}_i\}$ of the vector space of square matrices. Again we are taking familiar objects and looking deeper at their mathematical structure, which requires new notation like the tensor product to make explicit that structure.
\end{Example}
\begin{Example}\label{example:projections}\textbf{projections as linear transformations}
Trying to gain intuition about linear transformations $A: V\rightarrow V$ from a vector space into itself using the rotations and boosts of the plane is a bit misleading since they only give us intuition about linear transformations which are 1-1 and do not ``lose any points" as they move them around in the vector space on which they act. Such linear transformations are represented by nonsingular matrices when expressed in a basis, i.e., matrices with nonzero determinant $\det\ul{A}\neq 0$, which means that the only solution to
$\ul{A}\, \ul{x}=\ul{0}$ is the zero solution. Those matrices with zero determinant also arise naturally.
Suppose we decompose $V=V_1\oplus V_2$ into a direct sum of two subspaces, which simply means that any vector can be expressed uniquely as the sum of one vector in $V_1$ and another in $V_2$. In multivariable calculus, one of the first things we do with vector algebra is project a general vector in space into a vector component along a given direction and another one orthogonal to it. If $\hat u$ is a unit vector which picks out a direction in $\mathbb{R}^3$, then the projections of another vector $v$ parallel to and perpendicular to $\hat u$ are
$$
P_{u||}(v) = (v \cdot \hat u) \, \hat u\,,\quad
P_{u\bot}(v) = v- P_{u||}(v)
= v- (v \cdot \hat u) \, \hat u\,,
$$
If $v$ is already along $\hat u$, then the first projection just reproduces it, while the second gives the zero vector.
If $v$ is orthogonal to $\hat u$, then the second projection just reproduces it, while the first gives the zero vector.
By definition, the sum of the two projections just reproduces the original vector.
This is an example of a simple pair of projection maps $P$ and $Q$ which
satisfy $P^2=P$, $Q^2=Q$,$PQ=QP=0$ for a pair $(P,Q)$ which projects onto a pair of subspaces in a direct sum total space
$$
v \mapsto P(v)+Q(v) \,.
$$
Each acts as the identity on its corresponding subspace, and acts as the zero linear transformation on the other. This can be extended to a direct sum of any number of subspaces in an obvious way by iterating these conditions.
The vanishing of the determinant of a matrix $\ul{A}$ is the condition that the homogeneous linear system $\ul{A}\, \ul{x}=\ul{0}$ has nonzero solutions. The space of solutions is called either the null space or kernel of the matrix. Row reduction of the matrix produces a basis of that subspace of $\mathbb{R}^n$. However, there is no natural complementary subspace to complete projection into this subspace to a pair of projections as above without additional structure. The problem is that if $\ul{A}\, \ul{x} \neq\ul{0}$ then one can add any element of the null space to $\ul{x}$ and it will also satisfy the same condition of $\ul{x}$. But in $\mathbb{R}^n$ we have the orthogonal complement using the dot product to pick out a representative subspace we can use to decompose any vector into an element of the null space and another subspace. This is because the condition $A^i{}_j x^j=0$ means that the vector $\ul{x}$ is orthogonal to each of the rows of the coefficient matrix in the dot product interpretation, so that the span of the rows of the matrix (called the row space) is the orthogonal complement of the null space with this natural inner product. Similarly the set of all nonzero vectors $A^i{}_j x^j$ for all possible $x^j$ corresponds to what is called the range of the linear transformation, but by definition of span, this is simply the span of the set of columns of the matrix, called the column space of the matrix. This too has no natural complement without an inner product, but of course the dot product is ready to do the job. The row and column spaces of a matrix were discussed in detail in Section \ref{sec:VVast}.
\end{Example}
For every linear transformation $A : V \rightarrow V$, there is an associated linear transformation $A^T : V^* \rightarrow V^*$ called its transpose, defined by
$$
(A^T)(f)(u) = f(A(u)) = f_i A^i{}_j u^j
\rightarrow [(A^T)(f)]_j = f_i A^i{}_j\,,
$$
which takes the matrix form
$$
\ul{A^T(f)} =\ul{f}^T \ul{A}\,.
$$
Thus with the row vector $\ul{f}^T$ we associate the new row vector
$$
\ul{f}^T \mapsto \ul{f}^T \ul{A} \,,
$$
or equivalently taking the matrix transpose of this equation, the corresponding column vector $ \ul{f}$ is associated with the new column vector
$$
\ul{f} \mapsto \ul{A}^T \ul{f}\,.
$$
In words left multiplication of a row matrix $\ul{f}^T$ by the matrix $\ul{A}$ of the linear transformation $A$ is equivalent to right multiplication by the transpose matrix $\ul{A}^T$ of the corresponding transposed column matrix $\ul{f}$. The abstract transpose linear transformation therefore corresponds directly to the transposed matrix acting in the transposed direction by matrix multiplication. In other words, when we want to think of $V*$ as itself a vector space undergoing a linear transformation, we then want to think of its elements as column matrices multiplied on the left by a matrix, and this leads to the transpose matrix of the original transformation acting on $V$.
This transpose linear transformation corresponds to partial evaluation of the tensor $\mathbb{A}$ in its covector argument
$$
A^T: f \mapsto \mathbb{A}(f,\ )\,,
$$
resulting in a new covector. Thus the $(^1_1)$-tensor $\mathbb{A}$ packages both the linear transformation $A$ and its transpose $A^T$ in the same machine, so we can identify these particular transformations as two particular ways in which this tensor acts on both the vector space and its dual, and in fact we might as well use the same kernel symbol $A$ for the tensor as well, letting its partial evaluation in either argument represent the two respective linear transformations.
\begin{Example}\label{example:dotprod}\textbf{dot product as a tensor on $\mathbb{R}^n$}
The dot product of two vectors
$$
{\rm dot}(a,b)= a\cdot b = \langle a^1,\ldots,a^n\rangle \cdot \langle b^1,\ldots,b^n\rangle
= a^1 b^1 +\ldots + a^n b^n
= \dsum_{i=1}^n a^i b^i
= \ul{a}^T \ul{b}
$$
is the simplest bilinear scalar function of a pair of vectors in $\mathbb{R}^n$. It is therefore a $(^0_2)$-tensor ``${\rm dot}$."
The standard basis vectors are orthonormal so we need an appropriate symbol for their dot products, which will be interpreted as the components of the dot product tensor. Numerically the matrix of these components is the identity matrix but the index positioning must be covariant, so we introduce a covariant Kronecker delta symbol for these components
$$
{\rm dot}(e_i,e_j) = e_i \cdot e_j = \delta_{ij}
\equiv
\begin{cases}
1 & \text{if $i=j$\,,} \\
0 & \text{if $i\not=j$\,.}
\end{cases}
\,.
$$
Then by bilinearity
\begin{align*}
{\rm dot}(a,b)
&= a\cdot b % \langle a^1,\ldots,a^n\rangle \cdot \langle b^1,\ldots,b^n\rangle
= (a^i e_i)\cdot (b^j e_j) = a^i b^j e_i \cdot e_j = \delta_{ij}\, a^i b^j
& = \delta_{ij} \,\omega^i(a) \omega^j(b)
= \delta_{ij},\omega^i\otimes\omega^j(a,b)
\,,
\end{align*}
so this tensor is
$$
{\rm dot} = \delta_{ij}\omega^i\otimes\omega^j \,.
$$
The dot product on ${\mathbb R}^n$ is an example of an inner product on a vector space, named for its symbolic representation as a raised dot between the vector arguments.
An inner product on any vector space is a ``symmetric" ($^0_2$)-tensor which accepts two vector arguments in either order and produces a real number (and such that the determinant of its symmetric matrix of components is nonzero, the condition of nondegeneracy). The dot product on ${\mathbb R}^n$ is such an inner product whose matrix of components is the identity matrix with respect to the standard basis of ${\mathbb R}^n$. The index positioning $\delta_{ij}$ for a ($^0_2$) tensor shows that it is fundamentally different from the identity ($^1_1$)-tensor with components $\delta^i{}_j$, even though both matrices of components are the unit matrix. Section \ref{sec:LTV2Vast} will explore inner products on both $V$ and its dual space $V*$ and their interpretation as linear transformations.
\end{Example}
\subsection{More than 2 indices: general tensors}
If $T$ is a $(^p_q)$-tensor over $V$, then $T(f,g,\cdots,v,u,\cdots) \in \mathbb{R}$ is a scalar.
Define its components with respect to $\{e_i\}$ by
$$
T^{ij\cdots}_{\ \ mn\cdots} = T(\omega^i,\omega^j,\cdots,e_m,e_n,\cdots)
\qquad \text{(scalars)} \,.
$$
$p$ is the number of upper indices on these components, equal to the number of covector arguments, while $q$ is the number of lower indices, equal to the number of vector arguments, and it is convenient but not necessary to order all the covector arguments first and the vector arguments last.
Next introduce the $n^{p+q}$ basis ``vectors," i.e., ($^p_q$)-tensors
$$
\{
\underbrace{e_i\otimes e_j\otimes \cdots}_{p\rm\ factors}
\otimes
\underbrace{\omega^m\otimes\omega^n\otimes \cdots}_{q\rm\ factors}
\} \,.
$$
We can then expand any tensor in terms
this basis
$$
T=T^{ij\cdots}_{\ mn\cdots}
e_i\otimes e_j\otimes \cdots \otimes \omega^m\otimes\omega^n\otimes \cdots \,.
$$
This expansion follows from the multilinearity and the various definitions just as in the previous case of $(^1_1)$-tensors over $V$. Namely
$$
\meqalign{
T& (f,g,\ldots,u,v,\ldots) &&\\
&= T(f_i\omega^i,g_j\omega^j,\ldots,u^me_m, v^ne_n,\ldots)
&&\text{(argument component expansion)}
\\
&= f_ig_j\ldots u^m v^n\ldots T(\omega^i, \omega^j, \ldots, e_m, e_n\ldots)
&&\text{(multilinearity)},
\\
&= T^{ij\ldots}_{\ mn\ldots} f_ig_j\ldots u^m v^n\ldots
&&\text{(definition tensor components)}
\\
&= T^{ij\ldots}_{\ mn\ldots} e_i(f) e_j(g)\ldots \omega^m(u) \omega^n(v) \ldots
&&\text{(definition argument components)}
\\
&= (T^{ij\ldots}_{\ mn\ldots} e_i \otimes e_j \otimes\ldots \omega^m \otimes \omega^n\otimes \ldots) (f,g,\ldots,u,v,\ldots)
&&\,.\qquad \text{(definition tensor product)}
}
$$
Thus $T$ and its expansion in parentheses in the last line have the same value on any set of arguments, so they must be the same multilinear function.
\begin{Example}\label{example:tensorprods}\textbf{tensor products by multiplication}
The simplest tensor products are just multilinear functions of a set of vectors that result from multiplying
together in a certain order linear functions of a single vector.
For example, the product of the values of three linear functions of single vectors defines a multilinear function of three vectors by
$$
(f \otimes g\otimes h) (u,v,w)
= f(u) g(v) h(w) \,.
$$
Expressing this tensor in components leads to
$$
(f \otimes g\otimes h)
= (f \otimes g\otimes h)_{ijk} (\omega^i\otimes \omega^j\otimes \omega^k)
$$
where
$$
(f \otimes g\otimes h)_{ijk}= f_i g_j h_k\,.
$$
In other words we have constructed a $(^0_3)$-tensor $f \otimes g\otimes h$ from the tensor product of 3 covectors $f$, $g$, and $h$ and in terms of components in index notation, we have just multiplied their components together.
We can do the same thing with vectors instead of covectors
$$
u\otimes v\otimes w
= (u^i e_i)\otimes (v^j e_j)\otimes (w^k e_k)
= u^i v^j w^k e_i\otimes e_j\otimes e_k
\,,\quad
(u\otimes v\otimes w)^{ijk}
= u^i v^j w^k \,.
$$
Notice that
$$
(u\otimes v\otimes w)(f,g,h)=f(u) g(v) h(w) = (f \otimes g\otimes h) (u,v,w) \,.
$$
This is the same duality which allows us to think of a linear function of a single covector as a vector and vice versa, together sharing a natural pairing to produce the linear combination which is the value of the linear function.
\end{Example}
\begin{Example}\label{ex:determinant1}\textbf{determinant as a tensor}
On ${\mathbb R}^3$ with the usual dot and cross products, introduce the ($_3^0$)-tensor $D$ by
\begin{eqnarray}
D(u,v,w) = u \cdot (v \times w)
&=& \det
\left(
\begin{array}{ccc}
u^1 & u^2 & u^3 \\
v^1 & v^2 & v^3\\
w^1 & w^2 & w^3 \\
\end{array}
\right)
\nonumber\\
&=& \det
\left(
\begin{array}{ccc}
u^1 & v^1 & w^1 \\
u^2 & v^2 & w^2\\
u^3 & v^3 & w^3 \\
\end{array}
\right)
\qquad \text{(``triple scalar product")}
\nonumber\\
&=& \det \langle \ul{u} | \ul{v} | \ul{w} \rangle
\,,
\nonumber
\end{eqnarray}
where we use the property that the determinant is invariant under the transpose operation in order to keep our vectors associated with column matrices (while students usually see vectors as rows in the matrix in this context in calculus courses).
This is linear in each vector argument (the determinant is a linear function of each row or column, which should be obvious from its representation in terms of the linear dot and cross product operations). It therefore has the expansion
$$
D = D_{ijk}\omega^i\otimes\omega^j\otimes\omega^k \,,
$$
where
$$
D_{ijk} = D(e_i,e_j,e_k)
= e_i\cdot(e_j\times e_k)
=
\begin{cases}
1 & \mbox{if $(i,j,k)$ even perm.\ of (1,2,3)} \\
-1 & \mbox{if $(i,j,k)$ odd perm.\ of (1,2,3)} \\
0 & \mbox{otherwise} \\
\end{cases}
$$
so that
\begin{eqnarray}
D&=& \omega^1\otimes\omega^2\otimes\omega^3
+\omega^2\otimes\omega^3\otimes\omega^1
+\omega^3\otimes\omega^1\otimes\omega^2
\nonumber\\&&
-\omega^1\otimes\omega^3\otimes\omega^2
-\omega^2\otimes\omega^1\otimes\omega^3
-\omega^3\otimes\omega^2\otimes\omega^1 \,.
\nonumber
\end{eqnarray}
This corresponds directly to the usual explicit formula
$$
\det
\left(
\begin{array}{ccc}
u^1 & u^2 & u^3 \\
v^1 & v^2 & v^3\\
w^1 & w^2 & w^3 \\
\end{array}
\right)
= u^1 v^2 w^3 + u^2 v^3 w^1 + u^3 v^1 w^2 - u^1 v^3 w^2 - u^2 v^1 w^3 - u^3 v^2 w^1 \,.
$$
We will soon give this determinant component symbol $d_{ijk}$ a new name $\epsilon_{ijk}$ called the Levi-Civita symbol or Levi-Civita epsilon, since it is exactly what we need to handle the cross product in $\mathbb{R}^3$ and the easily prove vector identities involving the dot and cross-products, while generalizing to $\mathbb{R}^n$ to provide a terribly useful tool. In this new notation we then have
$$
\det \langle u | v | w \rangle = \epsilon_{ijk} \omega^i\otimes\omega^j\otimes\omega^k (u,v,w) = \epsilon_{ijk} u^i v^j w^k\,.
$$
\end{Example}
\begin{pro}\label{exercise:3x3det-cross-prod}\textbf{determinant as a tensor}
a)
Continuing the example, convince yourself that the nonzero components of the determinant function $\epsilon_{ijk}= e_i\cdot(e_j\times e_k)$ (which correspond directly to the 3 positive and 3 negative terms in the expansion of the determinant, respectively the 3 positive cyclic permutations of 123 and the 3 negative cyclic permutations of 123) are
$$
1=\epsilon_{123} = \epsilon_{231} = \epsilon_{312} = -\epsilon_{132} = -\epsilon_{213} = -\epsilon_{321} \,.
$$
b)
Notice that if we consider the determinant function $D(c,a,b)=\epsilon_{ijk} c^i a^j b^k$ unevaluated in its first (or last) vector input slot $D(\ ,a,b)=D(a,b,\ )$, we get one free index in the component representation of the resulting covector $f=D(\ ,a,b) = \epsilon_{ijk}a^jb^k \omega^i$. Show that this covector has components
$$
\langle f_1,f_2,f_3 \rangle = \langle a^2b^3-a^3b^2, a^3b^1-a^1b^3, a^1b^2-a^2b^1 \rangle = \langle (a\times b)^1, (a\times b)^2, (a\times b)^3 \rangle \,,
$$
which you recognize as the same components as the cross product vector $a\times b$. To make index position work out we must introduce a Kronecker delta with both indices up to write this in index form with our index conventions
$$
(a\times b)^i = \delta^{il}\epsilon_{ljk} a^j b^k\,.
$$
Thus we must introduce additional structure to understand this last shift in index position to take a covector to the corresponding vector with the same components. Let's wait till after the next exercise to start tackling that.
\end{pro}
\begin{pro}\label{exercise:quadprod}\textbf{quadruple scalar product}
On ${\mathbb R}^3$ with the usual dot and cross products, introduce the ($^0_4$)-tensor
$$
Q(u,v,w,z)=(u \times v)\cdot(w\times z)
$$
called the ``scalar quadruple product."
It satisfies an identity that we will prove easily in Chapter 4 \typeout{cross reference to chapter 4}%
$$
(a\times b) \cdot (c\times d) = (a\cdot c)(b\cdot d) -(a\cdot d)(b\cdot c)
= \left| \begin{matrix} a\cdot c & a\cdot d\\ b\cdot c & b\cdot d\end{matrix}\right| \,.
$$
Its components in the standard basis are
$Q_{ijmn}= (e_i\times e_j)\cdot (e_m\times e_n)$, from which one can immediately evaluate some of its nonzero components:
$Q_{2323}=Q_{3131}=Q_{1212}=1$. Check these values.
Notice that interchanging either $(i,j)$ or $(m,n)$ results in a sign change, but exchanging these pairs of indices does not
$$
Q_{ijkl} = -Q_{jikl}=-Q_{ijlk} = Q_{klij} \,.
$$
These ``symmetries" are important and will be explored below.
Show that this tensor satisfies one further cyclic identity (first index fixed, last 3 undergo a sum of all cyclic permutations)
$$
Q_{ijkl}+Q_{iklj}+Q_{iljk} = 0 \,.
$$
\end{pro}
The simplest example of a tensor created with the dot product is a covector:
$f_u(v)=u \cdot v$.
For each fixed $u$, this defines a linear function of $v$, i.e., a covector $f_u$. It is exactly this correspondence that allows one to avoid covectors in elementary linear algebra.
For a general inner product on any vector space, the degeneracy condition guarantees that this map from vectors to covectors is a vector space isomorphism and hence can be used to establish an identification between the vector space and its dual space. This will prove very useful. \typeout{Correction.}
\section{Tensor product and matrix multiplication}
By linearity, the components of the tensor product of a vector and a covector are
\begin{align*}
v\otimes f &=(v^i e_i) \otimes (f_j \omega^j)
&&\text{(expand in bases)}
\\
&= v^i f_j e_i \otimes \omega^j
&& \text{(factor out scalar coefficients)}
\\
&\equiv (v\otimes f)^i{}_j\, e_i \otimes \omega^j
&&\text{(definition of components of tensor)}
\\
&\rightarrow
(v\otimes f)^i{}_j=v^if_j &&
\end{align*}
or equivalently
$$
(v\otimes f)^i{}_j
=(v\otimes f)(\omega^i,e_j)
=\omega^i(v)f(e_i)
=v^i f_j \,.
$$
With the representation in component form of a vector and a covector as column and row matrices respectively, this tensor product is exactly equivalent to matrix multiplication
$$
\underline v \, \underline f^T
=\underbrace{\left(
\begin{array}{cc}
v^1 \\
\vdots \\
v^n \\
\end{array}
\right)}_{n\times 1} \underbrace{(f_1 \cdots f_n)}_{1\times n}
=
\underbrace{\left(
\begin{array}{cc}
v^1f_1 \ldots v^1f_n \\
\vdots \\
v^nf_1 \ldots v^nf_n \\
\end{array}
\right)}_{n\times n}
= (v^i f_j)\,,
$$
(the vector $\underline v$ is a column matrix, the covector $\underline f^T$ is a row matrix),
but in the opposite order from the evaluation of a covector on a vector, leading to a matrix rather than a scalar (number).
Thus matrix multiplication of a row matrix by a column matrix on the right represents the abstract evaluation operation of a covector on a vector or vice versa, while the matrix multiplication on the left represents the tensor product operation. In this sense the name ``scalar product"
for evaluation is more analogous to ``tensor product" (the first produces a ``scalar" or real number, the second a tensor).
\begin{Example}\label{example:dualbasisprojection}\textbf{dual basis vector projections}
When such a tensor product matrix product acts by matrix multiplication on a component vector on the right, it corresponds to evaluating the corresponding $({}^1_1)$-tensor on its second argument
$$
\ul{v}\, \ul{f}^T \, \ul{X} \leftrightarrow (v\otimes f) (\ ,X) = e_i v^i f_j X^j = f(X)\, v \,.
$$
This is exactly how the dual basis 1-forms project out the scalar components along their corresponding basis vectors
$ \omega^j(X) = X^j $. Multiplying the original basis vector by this scalar component yields the vector component along that basis vector $X^j e_j$ (no sum on $j$). The sum then recovers the original vector by adding all these separate vector components together.
For a new basis $e_{i'} = e^j{}_{i'} e_i =B^j{}_i e_j$, with corresponding dual basis
$\omega^{i'} = \omega^{i'}{}_j \omega^j = B^{-1 i}{}_j \omega^j$, the summed tensor product
$$
e_{i'}\otimes \omega^{i'} = B^j{}_i B^{-1i}{}_k e_j\otimes \omega^k
= \delta^j{}_k e_j\otimes \omega^k
= e_j\otimes \omega^j
$$
has the matrix representation
$$
\ul{e}_{i'} \, \ul{\omega}^{i'T}
= \ul{B} \, \ul{B}^{-1} = \ul{I}
$$
Each term in the matrix product sum over $i'$ is the projection matrix which picks out the {i'} vector component
$e_{i'}\otimes\omega^{i'}(\ ,X) = e_{i'}\omega^{i'}(X)=X^{i'} e_{i'}$ (no sum on $i'$)
of the component vector to which it is applied by matrix multiplication, i.e., by evaluation of the corresponding tensor product on its second argument.
\end{Example}
\begin{pro}\label{exercise:R2tensortransformation}\textbf{transforming a tensor on $\mathbb{R}^2$}
In Exercise \ref{exercise:R2changebasis}
the dual basis $W^1=\omega^1-\omega^2$, $W^2=-\omega^1+2\omega^2$ was found for the new basis
$\{E_1,E_2\}=\{\langle 2,1\rangle,\langle 1,1\rangle\} =\{ 2e_1+1e_2, 1e_1+1e_2\}$ on $\mathbb R^2$.
Find the components of the ($^1_1$)-tensor $T$ in terms of the standard basis $\{e_i\}$ and $\{\omega^i\}$
if $T$ has the following components in terms of the basis $\{E_i\}$:
$$
\begin{array}{ll}
T(W^1,E_1)=1\,,\ & T(W^1,E_2)=2\,,\\
T(W^2,E_1)=-1\,, & T(W^2,E_2)=0\,,
\end{array}
$$
i.e.
$$
T = 1 E_1\otimes W^1 + 2 E_1\otimes W^2
-1 E_2\otimes W^1 + 0 E_2\otimes W^2
= T^i{}_j\, e_i\otimes \omega^j
\,.
$$
Do this in two ways.
\\
a) Just substitute into $T$ the new basis vectors and dual basis covectors expressed in terms of the old ones and expand out the result to identify the 4 old components as the resulting coefficients of $e_i\otimes \omega^j$.
\\
b) Use the matrix transformation law that will be justified in the next section. With a prime introduced to distinguish components $T^{i'}{}_{j'}$ in the new basis $E_i=e_{i'}$ and dual basis $W^i = \omega^{i'}$ from those $T^i{}_j$ in the old basis, then the following matrix product will reproduce our previous result
$$
(T^{i'}{}_{j'})
=
\begin{pmatrix}
1 & 2\\
-1 & 0
\end{pmatrix} \,,\quad
\underline{T}' = \underline{A}\, \underline{T}\, \underline{A}^{-1}
\rightarrow
\underline{T}
=\underline{A}^{-1} \underline{T}'\, \underline{A}
= \underline{B}\, \underline{T}'\, \underline{B}^{-1}
\,,
$$
where the basis changing matrix and its inverse are
$$
\underline{B}^{-1} = \underline{A} =\begin{pmatrix} 1 & -1\\ -1 & 2\end{pmatrix}
= (W^i{}_j)
\,,\quad
\underline{B} = \underline{A}^{-1} =\begin{pmatrix} 2& 1\\ 1 & 1\end{pmatrix}
=(E^i{}_j)
\,.
$$
\end{pro}
% 2007-08-21 bob edit; remark1.txt file now included here
\begin{Remark}
Our notation is so compact that certain facts may escape us.
For example
$$
v\otimes f
= (v^i e_i)\otimes f
= v^i (e_i\otimes f)
= v^i e_i\otimes f
$$
is actually a distributive law for the tensor product. A simpler example shows this
$$
(u+v)\otimes f = u\otimes f+v\otimes f\,.
$$
How do we know this? Well, the only thing we know about the tensor product is how it is defined in terms of evaluation on its arguments
$$
\meqalign{
[(u+v)\otimes f](g,w)
&= g(u+v)f(w)
= [g(u)+g(v)]f(w)
\qquad &\text{(linearity)}
\nonumber\\
&=
g(u)f(w)+g(v)f(w)
\qquad &\text{(distributive law)}
\nonumber\\
&=
(u\otimes f)(g,w)+ (v\otimes f)(g,w)
\qquad &\text{({definition of $\otimes$})}
\nonumber\\
&=
[u\otimes f+ v\otimes f](g,w)
\qquad &\text{(linearity)}
}
$$
which is ``how one adds functions" to produce the sum function, namely by adding their values on arguments.
But if these functions inside the square brackets on each side of the equation have the same values on all pairs of arguments, they are the same function (i.e., ($^1_1$)-tensor), namely
$(u+v)\otimes f = u\otimes f + v\otimes f$.
In fact it is easy to show (exercise) that $(cv) \otimes f=c(v\otimes f)$ for any constant $c$, so in fact the tensor product behaves like a product should with linear combinations.
\end{Remark}
%\FigureHere
% figure 29
\begin{figure}[t]
\typeout{*** EPS figure 14}
\begin{center}
%\includegraphics[scale=0.2]{scan0014.ps}
\includegraphics[scale=0.4]{./figs/figactivetrans}
\end{center}
\caption{Active linear transformation of points: the point $v$ moves to the point $u=A(v)$.}
\label{fig:activetrans}
\end{figure}
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%% dg2:
% 1.5
\section{Linear transformations of $V$ into itself and a change of basis}
\label{sec:LTV2Vbasis}
%\input{LineartransformationsofVintoitselfandachangeofbasis.txt} %7
% 2007-09-26 bob edit
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
We have developed the description of a vector space $V$ and the tensor spaces that are defined ``over it" (first the dual space and then all of the $({}^p_q)$-tensor spaces as we add more and more upper and lower indices to the component symbols) starting from some fixed basis of $V$ which induces a basis of each of these other tensor spaces in terms of which we can express all tensors in terms of their components. However, the most interesting thing about this description is what happens to those components when we change the basis of $V$. This is accomplished via the space of $({}^1_1)$-tensors, which we have identified with the space of linear transformations of the vector space $V$ into itself. Partial evaluation of such a tensor on the vector argument leaves a vector value as the result---this is how one accomplishes a linear transformation: $u=u^i e_i\to A(\ ,u) = A^i{}_j e_i \,\omega^j(u)= A^i{}_j u^j e_i$.
Suppose $A : V\rightarrow V$ is a linear transformation of $V$ into itself.
If $\{e_i\}$ is a basis of $V$, then the matrix of $A$ with respect to $\{e_i\}$ is defined by
$$
\underline A = (A^i_{\mbox{ }j}) \,,\qquad
A^i{}_{j} = \omega^i(A(e_j)) = \text{$i$-th component of\ } A(e_j) \,.
$$
where
$i$ (left) is the row index and $j$ (right) the column index (first and second indices respectively, although the first is a superscript instead of the usual subscript like the second in the usual notation of elementary linear algebra). The $j$-th column of $\ul{A}$ is the column matrix $\underline{A(e_j)}$ of components of the vector $A(e_j)=A^i{}_j e_i$ with respect to the basis, denoted by underlining
$$
\underline{A}
=(\underline{A(e_1)}\
\underline{A(e_2)}\
\ldots\
\underline{A(e_n)}) \,.
$$
If we expand the equation
$$
u = A(v) \rightarrow
u^i e_i
= A(v^j e_j)
= v^j A(e_j)
= v^j A^i{}_j e_i
= (A^i{}_j v^j)e_i \,,
$$
we get the component relation $u^i = A^i{}_j v^j$
or its matrix form $\underline{u}=\underline{A}\, \underline{v}$,
where
$$
\underline{u}=
\left(
\begin{array}{c}
u^1 \\
\vdots \\
u^n \\
\end{array}
\right)
\,,\quad
\underline{v}=
\left(
\begin{array}{c}
v^1 \\
\vdots \\
v^n \\
\end{array}
\right)
$$
are the column matrices of components of $u$ and $v$ with respect to the basis.
\begin{figure}[h]
\typeout{*** EPS figure ??}
\begin{center}
%\includegraphics[scale=0.2]{scan0014.ps}
\includegraphics[scale=0.4]{./figs/figrotationgrids}
\end{center}
\caption{Active rotation of the plane showing the old and new bases and the old and new coordinate grids. Notice that the starting vector $X$ has the same relationship to the new axes as the vector $R^{-1}(X)$ rotated by the inverse rotation has to the original axes. In other words matrix multiplication by the inverse matrix gives the new components of a fixed vector with respect to the new rotated basis.}
\label{fig:rotationgrids}
\end{figure}
We can interpret this as an ``active" linear transformation of the points (vectors) of $V$ to new points of $V$. We start with a vector $v$ and end up at the new vector $u$ as shown in Fig.~\ref{fig:activetrans}. The rotation of the plane illustrated in Example 1.3.1 is a good example to keep in mind. Fig.~\ref{fig:rotationgrids} shows an active rotation of the standard basis and its grid by a $30^\circ$ rotation.
We can also use a linear transformation to change the basis of $V$, provided that it is nonsingular (its matrix has nonzero determinant), just the condition that the $n$ image vectors of the original basis $\{e_i\}$ are linearly independent so they can be used as a new basis. The point of view here is that general vectors do not move, but they change their components since they are expressed in terms of a new basis which is obtained by moving the old basis by the original active linear transformation. This mere change of coordinates is sometimes called a passive linear transformation since vectors remain fixed and are simply re-expressed in terms of a new basis which is obtained from the old basis by an active linear transformation.
If $B : V \rightarrow V$ is such a linear transformation, with matrix
$\underline {B} = (B^i{}_j) = (\omega^i(B(e_j))$ such that $\det \underline {B} \not =0$, then define $e_{i'} = B(e_i) = B^j{}_i e_j$.
As discussed above, the columns of
$\underline {B} = (\underline {B(e_1)} \cdots \underline {B(e_n)})$
are the components of a new basis vectors with respect to the old ones:
$B^i{}_j = \omega^i(B(e_j)) \equiv \omega^i(e_{j'})$ are the old components ($i$) of the $j$th new basis vector.
Primed indices will be associated with component expressions in the new basis.
Since $B$ is invertible, we have
$$
e_i
= B^{-1}(e_{i'})
= B^{-1j}{}_i e_{j'}\,,
$$
which states that the new components ($j$) of the old basis vectors ($i$) are the columns ($i$ fixed, $j$ variable) of the inverse matrix $\underline{B}^{-1}$.
The new basis $\{e_{i'}\}$ has its own dual basis $\{\omega^{i'}\}$ satisfying
$\omega^{i'}(e_{j'})=\delta^i{}_j$.
If we define
$$
\omega^{i'}=B^{-1i}{}_j \omega^{j}\,,
$$
which says that the rows of the inverse matrix ($i$ fixed, $j$ variable) are the old components of the new dual basis covectors,
then
\begin{eqnarray}
\omega^{i'}(e_{j'})
&=&
B^{-1i}{}_k \omega^k(B^{\ell}{}_j e_l)
= B^{-1i}{}_k B^{\ell}{}_j\delta^k{}_{\ell}
\nonumber\\
&=&
B^{-1i}{}_k B^k{}_j
= \delta^i{}_j \qquad (\mbox{since\ }
\underline{B}^{-1} \underline{B} = \underline{I})
\nonumber
\end{eqnarray}
confirms that this is the correct expression for the new dual basis.
Given any vector $v$, we can express it either in terms of the old basis or the new one
\begin{eqnarray}
&& v = v^i e_i \,,\
v^i = \omega^i(v)
\,,\nonumber\\
&& v = v^i{'} e_{i'}\,,\
v^{i'} = \omega^{i'}(v)
= B^{-1i}{}_j \omega^j(v)
= B^{-1i}{}_j v^j \,.
\nonumber
\end{eqnarray}
In other words, if we actively transform the old basis to a new basis using the linear transformation $B$, the new components of any vector are related to the old components of the \emph{same} vector by matrix multiplication by the inverse matrix $\underline B^{-1}$ as is clear from the rotation example in Fig.~\ref{fig:rotationgrids}
$$
\underline v'=\underline B^{-1}\underline{v}
$$
or equivalently
$$
\underline{v} = \underline{B}\, \underline{v'} \,.
$$
Similarly we can express any covector in terms of the old or new dual basis
\begin{eqnarray}
&& f = f_i \omega^i\,,\
f_i = f(e_i) \,,
\nonumber\\
&& f = f_{i'} \omega^{i'}\,,\
f_{i'} = f(e_{i'})
= f(B^j_{\ i} e_j)
= B^j_{\ i} f(e_j)
= f_j B^j{}_i \,,
\nonumber
\end{eqnarray}
i.e., the covector components transform by the matrix $\underline{B}$ but multiplying from the right if we represent covectors as row matrices
$$
(f_{1'}\cdot \cdot \cdot f_{n'})
= (f_{1}\cdot \cdot \cdot f_{n}) \underline {B}
\leftrightarrow
\underline {f}'{}^T = \underline {f}^T \underline {B}
$$
or equivalently
$$
\underline {f}^T = \underline {f'}^T \underline {B}^{-1} \,,
$$
where the explicit transpose makes it clear that $\underline {f}^T$ and $\underline {f}'{}^T$ are row vectors, necessary to multiply the square matrix on its left.
This describes a ``passive" transformation of $V$ into itself or of $V^*$ into itself, since the points of these spaces do not change but their components do change due to the change of basis.
Changing the basis actively by a linear transformation $B$ makes the components of vectors change by the inverse matrix $\underline {B}^{-1}$ of $B$, while an active transformation of $V$ into itself gives the components with respect to the unchanged basis of the new vectors as the matrix product by $\underline {B}$ with the old components. The active and passive transformations go in opposite directions so to speak.
%\FigureHere
% figure 15a
\begin{figure}[t]
\typeout{*** EPS figure 15a}
\begin{center}
%\includegraphics[scale=0.4]{./figs/figrotationbasis} \quad
%\includegraphics[scale=0.4]{./figs/figrotation}
\includegraphics[scale=0.4]{./figs/figrotationbasis} \quad
\includegraphics[scale=0.4]{./figs/figrotation}
\end{center}
\caption{
Left: the trigonometry of new basis vectors rotated by an angle $\theta$.
Right: a point $u$ can be rotated actively by the rotation to a new position
${B}\,{u}$ in terms of components with respect to the old basis, or it can simply be re-expressed passively in terms of the new rotated basis vectors, with new components
${u'}={B}^{-1}{u}$, which can be visualized by rotating $u$ in the opposite direction by the angle $\theta$ and expressing it with respect to the original basis vectors.
}
\label{fig:15a}
\end{figure}
\begin{Example}\label{example:rotcoordtransf}\textbf{rotation as a coordinate transformation}
Consider a rotation of the plane by an angle $\theta$, imagined as a small positive acute angle for purposes of illustration, see Fig.~\ref{fig:15a}. The basis vector $e_1=\langle 1,0\rangle$ is moved to the new basis vector $e_{1'}=\langle \cos\theta,\sin\theta\rangle$, while the basis vector $e_2=\langle 0,1\rangle$ is moved to the new basis vector $e_{2'}=\langle -\sin\theta,\cos\theta\rangle$ by the basic trigonometry shown in that figure, so the matrix whose columns are the new basis vectors is
$$
\underline{B}=\langle \underline{e}_{1'}|\underline{e}_{2'}\rangle
=\begin{pmatrix} \cos\theta & -\sin\theta\\ \sin\theta & \cos\theta \end{pmatrix}\,.
$$
Any point $\underline{u}=\langle u^1,u^2\rangle$ in the plane is rotated to its new position $\underline{B}\,\underline{u}$ as shown in the figure, but we can also re-express the same original vector $\ul{u}$ with respect to the new basis. Its new coordinates are related to its old coordinates by the inverse rotation
$$
\underline{u}' =\langle u^{1'}, u^{2'}\rangle
= \underline{B}^{-1} \underline{u} \,.
$$
\end{Example}
%\FigureHere
% figure 15
\begin{figure}[h]
\typeout{*** EPS figure 15}
\begin{center}
\includegraphics[scale=0.6]{./figs/scan0015}
\end{center}
\caption{
Left: active transformation, points move, basis fixed.
Right: passive transformation, points fixed, basis changes.
}
\label{fig:15}
\end{figure}
If we are more interested in merely changing bases than in active linear transformations, we can let
$A=B^{-1}$ so that the old components of vectors are multiplied by the matrix rather than the inverse matrix. Then we have
\begin{eqnarray}
\omega^{i'} &=& A^i{}_j\omega^j \longrightarrow v^{i'} = A^i{}_j v^j \,,
\nonumber\\
e_{i'} &=& A^{-1j}{}_{i} e_j \longrightarrow f_{i'} = f_j A^{-1j}{}_{i}\,,
\nonumber
\end{eqnarray}
Thus upper indices associated with vector component labels transform by the matrix $\underline {A}$ (whose rows are the old components of the new dual basis covectors), while lower indices associated with covector component labels transform by the matrix $\underline A^{-1}$ (whose columns are the old components of the new basis vectors).
In the jargon of this subject, these upper indices on components are called ``contravariant" while the lower indices on components called ``covariant". Vectors and covectors themselves are sometimes called ``contravariant vectors" and ``covariant vectors" respectively.
The above relations between old and new components of the same object are called ``transformation laws" for contravariant and covariant vector components.
By the linearity of the tensor product, these ``transformation laws" can extended to the components of any tensor.
For example, suppose $L=L^i{}_j \, e_i\otimes\omega^j$ is the $(^1_1)$-tensor associated with a linear transformation $L: V\longrightarrow V$, now using the same symbol for the linear transformation and the tensor.
Then
$$
\meqalign{
{L} &= L^i{}_j e_i\otimes\omega^j \,,\quad
L^i{}_j &= {L}(\omega^i,e_j) \,,
\\
{L} &= L^{i'}{}_{j'} e_{i'}\otimes\omega^{j'}\,,\
L^{i'}{}_{j'} &= {L}(\omega^{i'},e_{j'})
= {L}(A^i{}_k\omega^k\,,\
A^{-1\ell}{}_j e_{\ell})
= A^i{}_k A^{-1\ell}{}_j {L} (\omega^k,e_{\ell})
\\
& &= A^i{}_k A^{-1\ell}{}_j L^k{}_{\ell} \,.
}
$$
In other words the contravariant and covariant indices each transform by the appropriate factor of $A^i{}_j$ or $A^{-1i}{}_j$
$$
L^{i'}{}_{j'} = A^i{}_k A^{-1\ell}{}_{j} L^k{}_{\ell}
\quad \text{or inversely} \quad
L^{i}{}_{j} = A^{-1i}{}_k A^{\ell}{}_j L^{k'}{}_{\ell'} \,.
$$
This generalizes in an obvious way to any $(^p_q)$-tensor
$$
\meqalign{
T&=
T^{i\cdots}_{\ j\ldots} e_i\otimes \cdots\otimes\omega^j\otimes\cdots \,,\quad
T^{i\ldots}_{\ j\ldots}
&= T(\omega^i,\cdots,e_j,\cdots) \,,
\nonumber\\
T &=
T^{i'\ldots}_{\ j'\ldots} e_{i'}\otimes \cdots\otimes\omega^{j'}\otimes\cdots \,,\quad
T^{i'\ldots}_{\ j'\ldots}
&= T(\omega^{i'},\ldots,e_{j'},\ldots)
\nonumber\\
& &=
A^i{}_k \cdots A^{-1\ell}{}_{\ j}\cdots T^{k\cdots}_{\ \ell\cdots} \,.
}
$$
It is just a simple consequence of multilinearity.
\begin{Example}\label{example:transfidentity}\textbf{transforming the identity tensor}
We first defined the Knonecker delta just as a convenient shorthand symbol $\delta^i{}_j$, but then saw it coincided with the components of the evaluation or identity tensor
$$
I\!d
= \delta^i{}_j\mbox { }e_i\otimes\omega^j
= e_i\otimes\omega^i
= e_1\otimes\omega^1 + \cdots + e_n\otimes\omega^n\,.
$$
Since this must be true in any basis, if we ``transform" the Knonecker delta as the components of a $(^1_1)$-tensor, it should be left unchanged
$$
\delta^{i'}{}_{j'}
= A^i{}_k A^{-1\ell}{}_{\ j}\delta^k{}_{\ell}
= A^i{}_k A^{-1k}{}_{\ j}
= (\underline{A} \, \underline{A}^{-1})^i{}_j
= (\underline{I})^i{}_j
= \delta^i{}_j\,.
$$
The new components do equal the old!
\end{Example}
\subsection{Matrix form of the ``transformation law" for $(^1_1)$-tensors}
The ``transformation law" for the $(^1_1)$-tensor $ L$ associated with a linear transformation $L: V \longrightarrow V$ is
$$
L^{i'}{}_{j'} = A^{i}{}_k A^{-1\ell}{}_{j} L^k{}_{\ell}
= A^{i}{}_k L^k{}_{\ell} A^{-1\ell}{}_{j}
= [\underline{A}\, \underline{L}\, \underline{A}^{-1}]^i{}_j \,.
$$
In other words we recover the matrix transformation for a linear transformation under a change of basis discussed in the eigenvector problem
$$
\underline{L}' = \underline{A}\, \underline{L}\, \underline{A}^{-1}
$$
which leads to the conjugation operation (just a term for sandwiching a matrix between another matrix and its inverse), except that in the eigenvector change of basis discussion, this relation was written in terms of the inverse matrix $\underline{A}^{-1} = \underline{B}$
$$
\underline{L}'
= \underline{B}^{-1}\, \underline{L}\, \underline{B}
\longleftarrow
\text{columns of $B$ = old components of new basis vectors}
$$
which corresponds to the transformation of vector components
$$
\underline{v} = \underline{B}\, \underline{v}' \,,\qquad
\underline{v}' = \underline{B}^{-1} \underline{v}\,.
$$
Note that when one succeeds in finding a square matrix $\underline{B}=\langle \ul{b}_1| \ldots | \ul{b}_n\rangle$ of linearly independent eigenvectors of a matrix $\underline{L}$
(namely $\ul{L}\,\ul{b}_i = \lambda_i\, \ul{b}_i$
so that $\ul{L}\,\ul{B} = \langle \ul{L}\,\ul{b}_1| \ldots | \ul{L}\,\ul{b}_n\rangle
=\langle \lambda_1\,\ul{b}_1| \ldots | \lambda_n\,\ul{b}_n\rangle $), then the new components of the matrix with respect to a basis consisting of those eigenvectors is diagonal
$$
\underline{L}'
= \underline{B}^{-1}\, \underline{L}\, \underline{B}
= \begin{pmatrix}