Talk:Change of basis

Latest comment: 2 years ago by Anita5192 in topic Caption wrong for pic?

This article leaves much to be desired

edit

It is too obstruse and esoteric to be useful. This is a simple topic, can be presented with very simple examples in low-dimensional cases with basic linear algebra, and then more abstract and less relevant (to the majority of users) bits can then be given to expand on the basic simple points. Really what a tragedy when a beautiful topic is been presented in such a way as to elicit a gag reflex in 99/100 readers of the site. Truly frustrating. — Preceding unsigned comment added by 152.16.191.74 (talk) 07:59, 3 August 2015 (UTC)Reply

Something missing?

edit

Is there a phrase missing from the first sentence of https://en.wikipedia.org/wiki/Change_of_basis#Three_dimensions? --Ott0 (talk) 06:52, 2 November 2015 (UTC)Reply

Yes, I think. I added the missing words "let R". Dexo568 (talk) 22:37, 4 November 2015 (UTC)Reply

Preliminary notions statement seems cyclical

edit

In the second line of this section, namely: If T: RnRm is a linear transformation...

What then is the matrix of T(ej)? — Preceding unsigned comment added by 128.84.126.62 (talk) 09:01, 23 December 2016 (UTC)Reply

General case, derived using Clifford algebra

edit

A new section, "General case, derived using Clifford algebra," was recently inserted by Twy2008 (talk). I recommend removing it in its entirety, as it is far too complicated for this article and hence, not helpful.—Anita5192 (talk) 03:34, 27 January 2020 (UTC)Reply

Okay, I know how these kinds of talks go and they lead to removal. So, I removed it from the page. I have put a copy of it below for reference or in case someone would like to improve it and maybe re-add it in some revised form.Twy2008 (talk) 05:37, 27 January 2020 (UTC)Reply

General case, derived using Clifford algebra

edit

In Clifford algebra (or Geometric Algebra), we can write any vector   in an  -dimensional real Euclidean space   as  ,   (Einstein summation assumed), with scalar coefficients (coordinates)   on a standard orthonormal basis   with reciprocal basis  . On the standard basis  , the inner (dot) products are  , where   is the Kronecker delta and   is the   identity matrix. Generally, on any basis   with reciprocal basis  , we have  . Although it is possible to generalize   to  , the  -dimensional vector space over any field   with inner product, only   is considered in the following. A standard orthonormal basis   is always assumed to exist since one can always be found by using the Gram-Schmidt process. In fact, the linear (matrix) algebra representation of column vectors   and their dual row vectors   often assumes an orthonormal basis where   and   and  . However, the general case cannot assume an orthonormal basis and the inner product   on a general basis   is  . We can define other bases   for   with any sets of   linearly independent vectors as  ,  , etc., where a set of vectors   is linearly independent iff  , or iff the determinant of the matrix of basis vectors as columns is non-zero  . For any basis vector  , we can compute its reciprocal basis vector as   such that   or   (no summation) and  . Using basis vectors   and their reciprocal basis vectors  , the general case of change of basis, from any basis to the basis  , is expressed as  . Change of basis is a passive transformation that does not change the vector  , but only rewrites   in terms of another basis   or its reciprocal basis  . Change of basis to a basis  , etc., is again  , etc. Now, we must convert this Clifford algebra expression, for general change of basis of   from a basis   to a basis  , into matrix algebra forms. To do this, we must, in effect, convert from basis   back to the standard basis  , and then convert from standard basis   to basis  .

In linear algebra, the standard orthonormal basis  ,  , of   has each vector   represented as an  -element column vector   where the  th row is   and elsewhere  , and   is the transpose of matrix  . We can express vector   on the standard basis   as the column vector  . Then, we can express   on the basis   as the column vector   of coordinates on  . Each basis vector   is assumed to be in terms of the standard basis as  . Each reciprocal basis vector   can be expressed as  , where the coordinates   of   are  , which is an expression of Cramer's rule that constitutes a method of finding the inverse matrix   since  . The column vector  , on the standard basis  , is transformed onto the basis   as   and transformed back onto the standard basis   as  . The general case of the transformation of vector   from basis   to basis   is  , where   is the matrix of column vectors  ,  , and   is the inverse of the matrix of column vectors  . The transition matrix   for the general change of basis from the basis   to the basis   is therefore  . The matrix   transforms coordinates from the basis   to the standard basis  . The matrix   transforms coordinates from the standard basis   to the basis  . Finally,  , so that we can either compute an inverse matrix using linear algebra methods or, equivalently, compute a matrix of reciprocal row vectors using Clifford (geometric) algebra methods. Note that,   and   with Einstein summation over  .

In linear algebra, the dot product   of vector   expressed as the column vector   with coordinates on basis   and vector   expressed as the column vector   with coordinates on basis  , requires transforming the coordinates of   onto an arbitrary basis   as   and transforming the coordinates of   onto the reciprocal basis   as  . We can then form the dual of   on   as the row vector  , and finally the dot product is  . Therefore, the dot product is independent of the choice of basis  , or we may assume   is the standard basis  , where  . We may now write the dot product as  . If  , then   is called the Gram matrix. Furthermore, if  , then  .

General case, using tensor algebra

edit

Here is another proposed section. Like the other proposed section of mine, I guess that it also will not be wanted, but I offer it here for discussion. I think that this is better than the other proposal (which was about the general case using Clifford algebra and linear algebra). Maybe this general case using tensor algebra can be edited and included on the wiki someday. Change of basis is very important in tensor algebra and it seems like it should have a good discussion on this page. Twy2008 (talk) 12:18, 20 February 2020 (UTC)Reply

General case, using tensor algebra

edit

Introduction: Change of basis is fundamental to tensor algebra. The following presents a notation for change of basis using tensor algebra. It is primarily a notational problem to express tensor algebra clearly without seeming too complicated, but this is hard to avoid entirely and there is no exact standard for the notation. Anyone seeking to learn anything about advanced mathematics must accept and learn "complicated" notations sometimes. So please, try to bare with it, and try not to reject it immediately just because it seems "too complicated" for an encyclopedic article! The Einstein summation convention is used throughout.

Define the standard basis   for real ( )-dimensional pseudo-Euclidean space   with reciprocal basis   using the basis index set   [see Notation (indices), below]. Let   for  , and   for  . The covariant metric tensor for   is

 

and

 

where   is the Knocker delta. The contravariant metric tensor is  .

The vector   on   is  , and   on   is  . An index is lowered, and its vector is brought onto the reciprocal basis, as  . An index is raised onto the basis as  .

Now, define another general basis   for   with reciprocal basis   using the basis index set  .

Notation (indices): The superscript   on   is the part of the index name that indicates the basis of the index set. Index name   and index name   are completely different index names that run through   on basis   and   on basis  , independently (they would not indicate an index pair for contraction). The index names  , where   is any convenient integer (or omitted), are used to avoid the use of too many letters such as  ,  ,   etc. Therefore, index names with different values of the subscript   are also independent indices. The symbolic index name   has the numerical value   and can be abridged to   or   or   when the particular basis and/or vector do not need to be indicated, depending on the usage. Basically, the whole index name   is a value that runs from   to   which can also be expressed as   to mean the same thing. The contraction (or inner product) of two tensor indices, representing two vectors, cannot be performed until both indices are changed onto the same basis and also raised or lowered into a contra/co-variant pair. The contraction of any two basis vectors forms a metric tensor entry, possibly of mixed-bases. The contractions of a set of basis vectors with the same (or different) set of basis vectors forms the metric tensor (or the mixed-basis metric tensor between two different bases).

Let   and  , where

 

The last expression for   is a form of Cramer's rule, which is the  th cofactor   of   divided by the determinant  [1].

The covariant metric tensor on   is

 

The contravariant metric tensor on   is

 

The tensor   is the mixed-basis metric tensor for change of basis of contravariant coordinates  :

 

The tensor   is the mixed-basis metric tensor for change of basis of covariant coordinates  :

 

The inverse transformations are also possible since   when contracting one pair of matching indices while holding the other pair of matching indices to a fixed value, or   and  .

For references section:

Schouten, J. A. (1954), Ricci-Calculus (2nd ed.), Springer, doi:10.1007/978-3-662-12927-2

extra details

edit

For the expressions of  , more details are given below to show that they are equivalent. It is just a transposed matrix inverse, since the matrix of cofactors is not transposed as it usually is in linear algebra for the matrix inverse formula. This is all just an exercise and too much detail for the article, unless it were placed into a hiddle element that can pop open when clicked. Now, I'll stop cluttering up the Talk page any further unless there is some discussion!

 
@Twy2008 I think in editing these sort of articles, we should keep in mind that a majority of people visiting this page are likely students who are not entirely familiar with the topic. The article itself is already notation-heavy and somewhat pedantic. It abuses set notation to denote sequences. It introduces a lot of symbols. I think it's well-organized in general, but the "preliminary notions" section seems unnecessary and delays the main point. Consider my addition to the intro: For example, if   is a matrix whose columns comprise a basis of  , a vector   (in the standard basis) can also be expressed as a linear combination of  's columns by the vector  . By definition then,  . This summarizes the first two sections in two sentences. Can you summarize your addition without so much notation? AP295 (talk) 16:17, 13 January 2021 (UTC)Reply

References

  1. ^ Schouten (1954, p. 1)

Article needs improvement.

edit

The notation in this article needs some work. Sometimes it uses set notation for sequences, and sometimes it uses sequence-like notation for sets. I don't think there's any problem with defining "basis" as sequence of linearly independent spanning vectors rather than a set. This would also obviate the need for phrases like "ordered basis", which frequently appear in the article. Sequences are by definition ordered so editors should also avoid using redundant phrases like "ordered sequence".

Some of the examples also seem overly-complex. The example with Euler angles is unlikely to be helpful to a reader who's unfamiliar with the concept of a basis or change-of-basis. I like the organization though the content in the section "Preliminary notions" can be explained in-line with the rest of the content and does not need its own section in my opinion. Opinions? AP295 (talk) 14:25, 14 January 2021 (UTC)Reply

I tried to clean up some of this. Some of it was clearly redundant. However, the order of the vectors in a basis is important, and my textbooks use the phrases "ordered set" and "ordered basis" instead of "sequence," so I removed the word "sequence."—Anita5192 (talk) 17:31, 14 January 2021 (UTC)Reply
Thanks, though I think defining a "basis" as a sequence is more natural exactly because the order is important. In my book (that is, the book that I read), Linear Algebra Done Right, Axler uses "list", which is analogous to a sequence. A basis is linearly independent so by definition it contains no duplicate elements. Concatenating sequences is clearer on paper than taking the union of two bases, since the order of the union's elements would be ambiguous without any further explanation. Is there anything to be gained from calling them sets? AP295 (talk) 18:08, 14 January 2021 (UTC)Reply
@Anita5192 The article also uses an unfortunate mix of latex and some other sort of markup (e.g. "&alpha") for math, which makes it a bit difficult to edit. Looking it over again, I really think it needs to be changed to "sequence". For example, the "standard basis" is not just a set of vectors, it's a sequence. Using sequence notation makes things much simpler notation-wise, and better lends itself to inductive arguments and things of that nature. For example, if you have two bases (a1,a2,a3) and (b1,b2,b3) of two orthogonal subspaces of R^6, then (a1,a2,a3,b1,b2,b3) is a basis for R^6, and the order is unambiguous despite not being implied by the subscripts alone. As another example, in concatenating (a1,a2,a3) and (a1), the obtained sequence is not a basis and we simply say "the sequence is not a basis". However, the union of the sets {a1,a2,a3}, {a1} would be a basis, and we'd have to say something like awkward like "the elements of {a1,a2,a3} and {a1} are not a linearly independent collection of vectors". I'll do it at some point if that's okay with you. AP295 (talk) 15:16, 15 January 2021 (UTC)Reply
I am opposed to introducing the word "sequence," as that is not standard terminology in anything I have read. And I don't do much editing of mathematics typesetting, so I will leave that to someone else.—Anita5192 (talk) 16:49, 15 January 2021 (UTC)Reply
The textbook I learned from uses sequences. From a notational and linguistic standpoint, does it not make more sense to call them sequences for the reasons that I've stated? The term "sequence" is very common in analysis and many other subjects, so I don't understand your objection. AP295 (talk) 23:45, 15 January 2021 (UTC)Reply

This article did not state its main object, namely the basis-change formula. I have rewritten the article for fixing this, but most of the body still requires to be also rewritten. D.Lazard (talk) 19:59, 15 January 2021 (UTC)Reply

It does, but it's hard to pick out. Your changes have a somewhat editorial tone. I think this is too much for the intro, and I'm not sure why you deleted my example. I thought it was a pretty clear way of introducing the idea. You use the word "basis-change matrix" as if the reader is supposed to know what that means already, and it makes your definition a bit circular. My example demonstrated that changing from the standard basis to basis B is accomplished by multiplying with the inverse of B, and to the standard basis from basis B by multiplication with B itself. From there the reader understands that they can change from one arbitrary basis A to another B by taking (B^-1 A). Is this not much clearer and more explicit? And yes, the matrix of a linear map is determined by the map and a choice of bases for the domain and codomain, and I considered adding this to the intro myself, but I'm not sure if the intro is the place for it. It's an important thing to know but it is covered in the article itself. It just needs to be re-worked for clarity. AP295 (talk) 23:48, 15 January 2021 (UTC)Reply
@Anita5192, @D.Lazard Since the basis article also defines a basis as a set, I've opened up an RFC on that article's talk page. I believe this is an important distinction, and I invite you to comment if you have a legitimate objection to this and can explain it clearly. If you change your minds and agree, that would be just swell too. AP295 (talk) 01:43, 16 January 2021 (UTC)Reply
@D.Lazard In your comment on the maintenance tag, you wrote "most of the article is devoted to an overdetailed description of the fundamental concepts of linear algebra which are much better described in the relevant articles. This must be restricted to the minimum that is required for this article, and regrouped in a specific section". I agree with some of this but actually I think this article does a better job of introducing the concept of a "basis" than the actual article. If the notation is improved I think the content should be moved rather than deleted. AP295 (talk) 20:42, 16 January 2021 (UTC)Reply
I have fixed the indentation of the previous post.
When an article does not describe well its subject, it is against Wikipedia policy to rewrite it elsewhere (see WP:Content fork). So, it is a very bad idea to introduce here the concept of basis. In any case, edits do not delete anything, as anything that is removed can be rtrieved from the history of the article. D.Lazard (talk) 10:43, 17 January 2021 (UTC)Reply
"Moved" as in moved to Basis (linear algebra). AP295 (talk) 20:07, 20 January 2021 (UTC)Reply

I have rewritten the lead to be conform to MOS:MATH#Article introduction. As the article topic cannot be understand without a minimal knowledge of the subject, the lead must recall only what is relly needed to understand the phrasing. This is the reason for the removal of most of the previous lead. The new lead is incomplete, as it does not summarize really the content of the sections. In fact this cannot be done before fixing the flaws mentioned in the maintenance tags. D.Lazard (talk) 10:43, 17 January 2021 (UTC)Reply

Also, I have added the change-of-basis formula, because it is likely that many reader may come here to remember the exact form of the formula, in order to apply it correctly in their own application. So, it is important to give a direct access to this formula. D.Lazard (talk) 10:56, 17 January 2021 (UTC)Reply

Suggestion

edit

The ideal basis is the standard basis   which has the property

 

The cases where the inner product is zero mean that the vectors ei and ej are orthogonal to each other, while the cases   mean that the vectors are unit vectors. The standard basis is said to be orthonormal, meaning that the above property holds.

Now if   is a set of linearly independent vectors, then the matrix M with these vectors as rows is an invertible matrix and has the property   for each i = 1,...,n.

To show that   is a basis an arbitrary w in the vector space must be expressible in that basis. Let   and express v in the standard basis:   Then

 

One reason to make a change of basis may be to take a set of linearly independent vectors like {fi} (above) and use it to construct an orthonormal basis. As indicated, this procedure is equivalent to finding the inverse of the matrix using the f’s as rows. The step-wise procedure uses elementary rotations and provides Gram-Schmidt orthogonalization.

Another example of a useful change of basis re-configures rectangular hyperbolas to establish the hyperbolic functions sinh and cosh. Change of basis is also used in abstract algebra to relate alternate representations of algebras.

Perhaps the article would improve with this approach. Rgdboer (talk) 03:38, 11 February 2021 (UTC)Reply

How could this exist so long and nobody noticed that "basis" is used, but not defined, in the lede?

edit

Well? Huh? Well?

VerdanaBold 12:20, 9 July 2021 (UTC)Reply

A link is provided for a definition, and a characteristic property is given. This should suffice for people who have already heard of a basis. For others, this is unimportant, as it is irrealistic to try changing something that one does not know about. D.Lazard (talk) 14:32, 9 July 2021 (UTC)Reply

Change of basis formulas are wrong

edit

The change of basis formulas for linear maps, endomorphisms, and bilinear forms are wrong. For instance, for endomorphisms, it should be   and not  . Right? Seub (talk) 00:50, 18 August 2021 (UTC)Reply

  Fixed. Good point. A way to remember this formula is that the product of matrices must provide "new coordintes". So the matrix on the left must express "new" coordinates in terms of "old" ones. So, it must be the inverse of the change-of-basis matrix, which expresses the "old" coordintes in terms of the "new" ones. D.Lazard (talk) 07:56, 18 August 2021 (UTC)Reply
Thanks for the correction. The new matrix of the bilinear form also needs to be changed: it is   and not  . Seub (talk) 10:58, 18 August 2021 (UTC)Reply
Seub,   Done. You could have done this by yourself. D.Lazard (talk) 14:27, 18 August 2021 (UTC)Reply

Caption wrong for pic?

edit

the second illustration picture, when you click it to view by itself, displays the caption "A vector (here in 3d, shown the purple arrow) can be represented in terms of two different bases (green and blue arrows), each basis vector is scalar-multiplied appropriately so they add to the vector."

But there are no green or blue items in that pic. 2601:80:4300:3AE0:74D3:9ED1:B92F:79C4 (talk) 03:07, 19 October 2022 (UTC)Reply

I cannot display this caption. The only caption that I have found is "A vector represented by two different bases (purple and red arrows)". D.Lazard (talk) 08:33, 19 October 2022 (UTC)Reply
I can see what the IP is referring to. When I click the image in the article, I am taken to the image. When I scroll down, I see the erroneous description.—Anita5192 (talk) 15:52, 19 October 2022 (UTC)Reply