This article is within the scope of WikiProject Mathematics, a collaborative effort to improve the coverage of mathematics on Wikipedia. If you would like to participate, please visit the project page, where you can join the discussion and see a list of open tasks.MathematicsWikipedia:WikiProject MathematicsTemplate:WikiProject Mathematicsmathematics articles
Latest comment: 9 years ago1 comment1 person in discussion
It is too obstruse and esoteric to be useful. This is a simple topic, can be presented with very simple examples in low-dimensional cases with basic linear algebra, and then more abstract and less relevant (to the majority of users) bits can then be given to expand on the basic simple points. Really what a tragedy when a beautiful topic is been presented in such a way as to elicit a gag reflex in 99/100 readers of the site. Truly frustrating. — Preceding unsigned comment added by 152.16.191.74 (talk) 07:59, 3 August 2015 (UTC)Reply
Latest comment: 4 years ago2 comments2 people in discussion
A new section, "General case, derived using Clifford algebra," was recently inserted by Twy2008 (talk). I recommend removing it in its entirety, as it is far too complicated for this article and hence, not helpful.—Anita5192 (talk) 03:34, 27 January 2020 (UTC)Reply
Okay, I know how these kinds of talks go and they lead to removal. So, I removed it from the page. I have put a copy of it below for reference or in case someone would like to improve it and maybe re-add it in some revised form.Twy2008 (talk) 05:37, 27 January 2020 (UTC)Reply
In Clifford algebra (or Geometric Algebra), we can write any vector
in an -dimensional real Euclidean space as
, (Einstein
summation assumed), with scalar coefficients (coordinates) on a
standard orthonormal basis with reciprocal basis . On the standard basis, the inner (dot)
products are ,
where is the Kronecker delta and is the identity matrix. Generally, on any basis with
reciprocal basis , we have . Although it is possible to generalize
to , the -dimensional vector space over any
field with inner product, only is considered in
the following. A standard orthonormal basis is always
assumed to exist since one can always be found by using the Gram-Schmidt process.
In fact, the linear (matrix) algebra representation of column vectors
and their dual row vectors often assumes an orthonormal basis
where and and . However, the general case cannot assume an orthonormal basis and the inner product on a general basis
is . We can
define other bases for with any sets of
linearly independent vectors as ,
, etc., where a set of vectors is linearly independent iff , or iff the
determinant of the matrix of basis vectors as columns is non-zero . For any basis vector
, we can compute its reciprocal basis vector as such that or (no summation) and . Using basis vectors
and their reciprocal basis vectors , the general
case of change of basis, from any basis to the basis , is expressed as
.
Change of basis is a passive transformation that does not change the vector
, but only rewrites in terms of another basis
or its reciprocal basis. Change of basis to a
basis , etc., is again , etc. Now, we
must convert this Clifford algebra expression, for general change of basis of
from a basis to a basis , into matrix algebra forms. To do this, we must, in effect,
convert from basis back to the standard basis, and then convert from
standard basis to basis .
In linear algebra, the standard orthonormal basis, , of has each vector
represented as an -element column vector where the th row is
and elsewhere , and is the transpose of matrix. We can express vector on the standard basis as
the column vector .
Then, we can express on
the basis as the column vector of coordinates on . Each basis vector
is assumed to be in terms of the standard basis as
. Each reciprocal basis vector
can be expressed as ,
where the coordinates of are , which is an expression of Cramer's rule that constitutes a
method of finding the inverse matrix since . The column vector , on the
standard basis, is transformed onto the basis as and transformed back onto the standard basis as . The general case of the transformation of vector
from basis to basis is , where is the matrix of column vectors , , and
is the inverse of the matrix of column vectors . The transition
matrix for the general change of basis from the basis to the basis is therefore . The matrix transforms coordinates from the basis to the
standard basis. The matrix transforms
coordinates from the standard basis to the basis . Finally, , so that we can either
compute an inverse matrix using linear algebra methods or, equivalently,
compute a matrix of reciprocal row vectors using Clifford (geometric) algebra
methods. Note that, and
with Einstein summation over .
In linear algebra, the dot product of vector expressed as the column vector with
coordinates on basis and vector expressed as the column vector with coordinates on basis ,
requires transforming the coordinates of
onto an arbitrary basis as and transforming the coordinates of
onto the reciprocal basis as . We can then form the dual of
on as the row vector , and finally the dot product is
. Therefore, the dot product is independent of
the choice of basis , or we may assume is the standard basis, where . We may now write the
dot product as . If , then is called the Gram matrix. Furthermore, if
, then .
Latest comment: 3 years ago2 comments2 people in discussion
Here is another proposed section. Like the other proposed section of mine, I guess that it also will not be wanted, but I offer it here for discussion. I think that this is better than the other proposal (which was about the general case using Clifford algebra and linear algebra). Maybe this general case using tensor algebra can be edited and included on the wiki someday. Change of basis is very important in tensor algebra and it seems like it should have a good discussion on this page. Twy2008 (talk) 12:18, 20 February 2020 (UTC)Reply
Introduction: Change of basis is fundamental to tensor algebra. The following
presents a notation for change of basis using tensor algebra. It is primarily
a notational problem to express tensor algebra clearly without seeming too
complicated, but this is hard to avoid entirely and there is no exact standard
for the notation. Anyone seeking to learn anything about advanced mathematics
must accept and learn "complicated" notations sometimes. So please, try to
bare with it, and try not to reject it immediately just because it seems "too
complicated" for an encyclopedic article! The Einstein summation convention
is used throughout.
Define the standard basis for real ()-dimensional pseudo-Euclidean space with reciprocal
basis using the basis index set [see Notation
(indices), below]. Let for , and for . The covariant metric tensor for is
and
where is the Knocker delta. The contravariant metric tensor is
.
The vector on is , and
on is . An
index is lowered, and its vector is brought onto the reciprocal basis, as
. An index is raised onto the basis as .
Now, define another general basis for
with reciprocal basis using the
basis index set .
Notation (indices): The superscript on is the part of the index
name that indicates the basis of the index set. Index name and index
name are completely different index names that run through on basis and on basis ,
independently (they would not indicate an index pair for contraction). The
index names , where is any convenient integer (or omitted), are used
to avoid the use of too many letters such as , ,
etc. Therefore, index names with different values of the subscript are
also independent indices. The symbolic index name has the numerical
value and can be abridged to or or when the particular
basis and/or vector do not need to be indicated, depending on the usage.
Basically, the whole index name is a value that runs from
to which can also be expressed as to mean the
same thing. The contraction (or inner product) of two tensor indices,
representing two vectors, cannot be performed until both indices are changed
onto the same basis and also raised or lowered into a contra/co-variant pair.
The contraction of any two basis vectors forms a metric tensor entry, possibly
of mixed-bases. The contractions of a set of basis vectors with the same (or
different) set of basis vectors forms the metric tensor (or the mixed-basis
metric tensor between two different bases).
Let and , where
The last expression for is a form of Cramer's
rule, which is the th cofactor of
divided by the
determinant [1].
The covariant metric tensor on is
The contravariant metric tensor on is
The tensor is the mixed-basis metric tensor for change of
basis of contravariant coordinates :
The tensor is the mixed-basis metric tensor for change of
basis of covariant coordinates :
The inverse transformations are also possible since when contracting one pair of matching indices
while holding the other pair of matching indices to a fixed value, or
and
.
For the expressions of , more details are given below to show that they are equivalent. It is just a transposed matrix inverse, since the matrix of cofactors is not transposed as it usually is in linear algebra for the matrix inverse formula. This is all just an exercise and too much detail for the article, unless it were placed into a hiddle element that can pop open when clicked. Now, I'll stop cluttering up the Talk page any further unless there is some discussion!
@Twy2008 I think in editing these sort of articles, we should keep in mind that a majority of people visiting this page are likely students who are not entirely familiar with the topic. The article itself is already notation-heavy and somewhat pedantic. It abuses set notation to denote sequences. It introduces a lot of symbols. I think it's well-organized in general, but the "preliminary notions" section seems unnecessary and delays the main point. Consider my addition to the intro: For example, if is a matrix whose columns comprise a basis of , a vector (in the standard basis) can also be expressed as a linear combination of 's columns by the vector . By definition then, . This summarizes the first two sections in two sentences. Can you summarize your addition without so much notation? AP295 (talk) 16:17, 13 January 2021 (UTC)Reply
Latest comment: 3 years ago14 comments3 people in discussion
The notation in this article needs some work. Sometimes it uses set notation for sequences, and sometimes it uses sequence-like notation for sets. I don't think there's any problem with defining "basis" as sequence of linearly independent spanning vectors rather than a set. This would also obviate the need for phrases like "ordered basis", which frequently appear in the article. Sequences are by definition ordered so editors should also avoid using redundant phrases like "ordered sequence".
Some of the examples also seem overly-complex. The example with Euler angles is unlikely to be helpful to a reader who's unfamiliar with the concept of a basis or change-of-basis. I like the organization though the content in the section "Preliminary notions" can be explained in-line with the rest of the content and does not need its own section in my opinion. Opinions? AP295 (talk) 14:25, 14 January 2021 (UTC)Reply
I tried to clean up some of this. Some of it was clearly redundant. However, the order of the vectors in a basis is important, and my textbooks use the phrases "ordered set" and "ordered basis" instead of "sequence," so I removed the word "sequence."—Anita5192 (talk) 17:31, 14 January 2021 (UTC)Reply
Thanks, though I think defining a "basis" as a sequence is more natural exactly because the order is important. In my book (that is, the book that I read), Linear Algebra Done Right, Axler uses "list", which is analogous to a sequence. A basis is linearly independent so by definition it contains no duplicate elements. Concatenating sequences is clearer on paper than taking the union of two bases, since the order of the union's elements would be ambiguous without any further explanation. Is there anything to be gained from calling them sets? AP295 (talk) 18:08, 14 January 2021 (UTC)Reply
@Anita5192 The article also uses an unfortunate mix of latex and some other sort of markup (e.g. "&alpha") for math, which makes it a bit difficult to edit. Looking it over again, I really think it needs to be changed to "sequence". For example, the "standard basis" is not just a set of vectors, it's a sequence. Using sequence notation makes things much simpler notation-wise, and better lends itself to inductive arguments and things of that nature. For example, if you have two bases (a1,a2,a3) and (b1,b2,b3) of two orthogonal subspaces of R^6, then (a1,a2,a3,b1,b2,b3) is a basis for R^6, and the order is unambiguous despite not being implied by the subscripts alone. As another example, in concatenating (a1,a2,a3) and (a1), the obtained sequence is not a basis and we simply say "the sequence is not a basis". However, the union of the sets {a1,a2,a3}, {a1} would be a basis, and we'd have to say something like awkward like "the elements of {a1,a2,a3} and {a1} are not a linearly independent collection of vectors". I'll do it at some point if that's okay with you. AP295 (talk) 15:16, 15 January 2021 (UTC)Reply
I am opposed to introducing the word "sequence," as that is not standard terminology in anything I have read. And I don't do much editing of mathematics typesetting, so I will leave that to someone else.—Anita5192 (talk) 16:49, 15 January 2021 (UTC)Reply
The textbook I learned from uses sequences. From a notational and linguistic standpoint, does it not make more sense to call them sequences for the reasons that I've stated? The term "sequence" is very common in analysis and many other subjects, so I don't understand your objection. AP295 (talk) 23:45, 15 January 2021 (UTC)Reply
This article did not state its main object, namely the basis-change formula. I have rewritten the article for fixing this, but most of the body still requires to be also rewritten. D.Lazard (talk) 19:59, 15 January 2021 (UTC)Reply
It does, but it's hard to pick out. Your changes have a somewhat editorial tone. I think this is too much for the intro, and I'm not sure why you deleted my example. I thought it was a pretty clear way of introducing the idea. You use the word "basis-change matrix" as if the reader is supposed to know what that means already, and it makes your definition a bit circular. My example demonstrated that changing from the standard basis to basis B is accomplished by multiplying with the inverse of B, and to the standard basis from basis B by multiplication with B itself. From there the reader understands that they can change from one arbitrary basis A to another B by taking (B^-1 A). Is this not much clearer and more explicit? And yes, the matrix of a linear map is determined by the map and a choice of bases for the domain and codomain, and I considered adding this to the intro myself, but I'm not sure if the intro is the place for it. It's an important thing to know but it is covered in the article itself. It just needs to be re-worked for clarity. AP295 (talk) 23:48, 15 January 2021 (UTC)Reply
@Anita5192, @D.Lazard Since the basis article also defines a basis as a set, I've opened up an RFC on that article's talk page. I believe this is an important distinction, and I invite you to comment if you have a legitimate objection to this and can explain it clearly. If you change your minds and agree, that would be just swell too. AP295 (talk) 01:43, 16 January 2021 (UTC)Reply
@D.Lazard In your comment on the maintenance tag, you wrote "most of the article is devoted to an overdetailed description of the fundamental concepts of linear algebra which are much better described in the relevant articles. This must be restricted to the minimum that is required for this article, and regrouped in a specific section". I agree with some of this but actually I think this article does a better job of introducing the concept of a "basis" than the actual article. If the notation is improved I think the content should be moved rather than deleted. AP295 (talk) 20:42, 16 January 2021 (UTC)Reply
I have fixed the indentation of the previous post.
When an article does not describe well its subject, it is against Wikipedia policy to rewrite it elsewhere (see WP:Content fork). So, it is a very bad idea to introduce here the concept of basis. In any case, edits do not delete anything, as anything that is removed can be rtrieved from the history of the article. D.Lazard (talk) 10:43, 17 January 2021 (UTC)Reply
I have rewritten the lead to be conform to MOS:MATH#Article introduction. As the article topic cannot be understand without a minimal knowledge of the subject, the lead must recall only what is relly needed to understand the phrasing. This is the reason for the removal of most of the previous lead. The new lead is incomplete, as it does not summarize really the content of the sections. In fact this cannot be done before fixing the flaws mentioned in the maintenance tags. D.Lazard (talk) 10:43, 17 January 2021 (UTC)Reply
Also, I have added the change-of-basis formula, because it is likely that many reader may come here to remember the exact form of the formula, in order to apply it correctly in their own application. So, it is important to give a direct access to this formula. D.Lazard (talk) 10:56, 17 January 2021 (UTC)Reply
Latest comment: 3 years ago1 comment1 person in discussion
The ideal basis is the standard basis which has the property
The cases where the inner product is zero mean that the vectors ei and ej are orthogonal to each other, while the cases mean that the vectors are unit vectors. The standard basis is said to be orthonormal, meaning that the above property holds.
Now if is a set of linearly independent vectors, then the matrix M with these vectors as rows is an invertible matrix and has the property for each i = 1,...,n.
To show that is a basis an arbitrary w in the vector space must be expressible in that basis. Let and express v in the standard basis: Then
One reason to make a change of basis may be to take a set of linearly independent vectors like {fi} (above) and use it to construct an orthonormal basis. As indicated, this procedure is equivalent to finding the inverse of the matrix using the f’s as rows. The step-wise procedure uses elementary rotations and provides Gram-Schmidt orthogonalization.
Another example of a useful change of basis re-configures rectangular hyperbolas to establish the hyperbolic functions sinh and cosh. Change of basis is also used in abstract algebra to relate alternate representations of algebras.
A link is provided for a definition, and a characteristic property is given. This should suffice for people who have already heard of a basis. For others, this is unimportant, as it is irrealistic to try changing something that one does not know about. D.Lazard (talk) 14:32, 9 July 2021 (UTC)Reply
Latest comment: 3 years ago4 comments2 people in discussion
The change of basis formulas for linear maps, endomorphisms, and bilinear forms are wrong. For instance, for endomorphisms, it should be and not . Right? Seub (talk) 00:50, 18 August 2021 (UTC)Reply
Fixed. Good point. A way to remember this formula is that the product of matrices must provide "new coordintes". So the matrix on the left must express "new" coordinates in terms of "old" ones. So, it must be the inverse of the change-of-basis matrix, which expresses the "old" coordintes in terms of the "new" ones. D.Lazard (talk) 07:56, 18 August 2021 (UTC)Reply
Latest comment: 2 years ago3 comments3 people in discussion
the second illustration picture, when you click it to view by itself, displays the caption "A vector (here in 3d, shown the purple arrow) can be represented in terms of two different bases (green and blue arrows), each basis vector is scalar-multiplied appropriately so they add to the vector."
I cannot display this caption. The only caption that I have found is "A vector represented by two different bases (purple and red arrows)". D.Lazard (talk) 08:33, 19 October 2022 (UTC)Reply
I can see what the IP is referring to. When I click the image in the article, I am taken to the image. When I scroll down, I see the erroneous description.—Anita5192 (talk) 15:52, 19 October 2022 (UTC)Reply