Note = edit

Another note -- should mention that the generalization of the notion of transpose to complex matrices is to make element (i,j) equal to the conjugate of the (j,i) element. I agree with Chris W that the notation A' should be mentioned as an alternative. Wile E. Heresiarch 08:39, 11 Mar 2004 (UTC)

Differentiation of transposed matrices edit

I was trying to understand what the derivative of a transposed matrix is with respect to that matrix? So something like   where both are matrices.
yanneman 13:26, 20 November 2006 (UTC)Reply

Transpose of linear maps edit

In the Transpose of linear maps section, the article had read:

If f : V→W is a linear map between vector spaces V and W with dual spaces W* and V*, we define the transpose of f to be the linear map tf : W*→V* with
 
  for every   in W*.

An anonymous user changed that to

If f : V→W is a linear map between vector spaces V and W with dual spaces W* and V*, we define the transpose of f to be the linear map fT : W*→V* with
 
  for every   in W*.

Now, I think that's confusing (it's f that's transposed, not f(φ)), and it would be better to write

 
  for every   in W*.

But even that seems ambiguous; is this "transpose f" or "f to the T power"? But I'm unfamiliar with this notation, and I certainly don't have the same objection to T used with matrices. So, are there any experts who could weigh in with the most common usage in this area? (I'm not proposing to change the T notation used in the earlier sections, just to keep the t notation in this one section.) --Quuxplusone 21:14, 8 August 2005 (UTC)Reply

Putting the T in front sounds like a worthy experiment. In general though, you have to realise that mathematics notation is never anywhere near as rigorous as people would like to think. A classic example is -1 designates both the inverse and the reciprocal. Cesiumfrog (talk) 04:52, 11 May 2010 (UTC)Reply
What about using  ? —Ben FrantzDale (talk) 12:40, 11 May 2010 (UTC)Reply

hermitian transpose? edit

hermitian transpose = conjugate transpose? --Moala 09:23, 20 December 2005 (UTC)

Transpose on tensors... edit

I'm confused. It seems like much of linear algebra glosses over the meaning of transposition and simply uses it as a mechanism for manipulating numbers, for example, defining the norm of v as  .

In some linear-algebra topics, however, it appears that column and row vectors have different meanings (that appear to have something to do with covariance and contravariance of vectors). The transpose of a column vector, c, gives you a row vector -- a vector in the dual space of c. I think the idea is that column vectors would be indexed with raised indices and row vectors with lowered indices with tensors.

Here's my confusion: If row vectors and column vectors are in distinct spaces (and they certainly are in that you can't just add them), then taking the transpose of a vector isn't just some notational convenience, it is an application of a nontrivial function,  . To do something like this in general, we can use any bilinear form, but that involves more structure than just

So:

  1. Is it correct that there are are two things going on here: (1) using transpose for numerical convenience and (2) using rows versus columns to for indicateng co- versus contravariance?
  2. Isn't the conventional Euclidean metric defined with a contravariant metric tensor:  ? Doesn't that not involve any transposition in that both vs have raised indices?

Thanks. —Ben FrantzDale (talk) 05:00, 11 November 2009 (UTC)Reply

As asked on the Math Reference Desk edit

Transpose and tensors edit

I posed a question on Talk:Transpose that didn't get any responses there. Perhaps this is a better audience since it's a bit of an essoteric question for such an elementary topic; Here's the question again:

I'm confused. It seems like much of linear algebra glosses over the meaning of transposition and simply uses it as a mechanism for manipulating numbers, for example, defining the norm of v as  .
In some linear-algebra topics, however, it appears that column and row vectors have different meanings (that appear to have something to do with covariance and contravariance of vectors). In that context, the transpose of a column vector, c, gives you a row vector -- a vector in the dual space of c. I think the idea is that column vectors would be indexed with raised indices and row vectors with lowered indices with tensors.
Here's my confusion: If row vectors and column vectors are in distinct spaces (and they certainly, even in elementary linear algebra in that you can't just add a column to a row vector because they have different shapes), then taking the transpose of a vector isn't just some notational convenience, it is an application of a nontrivial function,  . To do something like this in general, we can use any bilinear form, but that involves more structure than just
So:
  1. Is it correct that there are are two things going on here: (1) using transpose for numerical convenience and (2) using rows versus columns to for indicateng co- versus contravariance?
  2. Isn't the conventional Euclidean metric defined with a contravariant metric tensor:  ? Doesn't that not involve any transposition in that both vs have raised indices?
Thanks. —Ben FrantzDale (talk) 14:16, 23 November 2009 (UTC)Reply
I guess it depends on how we define vectors. If we consider a vector as just being an n by m matrix with either n=1 or m=1, then transposition is just what it is with any other matrix - a map from the space of n by m matrices to the space of m by n matrices. --Tango (talk) 14:38, 23 November 2009 (UTC)Reply
Sure. I'm asking because I get the sense that there are some unwritten ruels going on. At one extreme is the purely-mechanical notion of tranpose that you describe, which I'm happy with. In that context, transpose is just used along with matrix operations to simplify the expression of some operations. At the other extreme, rows and columns correspond to co- and contra-variant vectors, in which case transpose is completely non-trivial.
My hunch is that the co- and contravariance convention is useful for some limited cases in which all transformations are mixex of type (1,1) and all (co-) vectors are either of type (0,1) or (1,0). But that usage doesn't extend to problems involving things like type-(0,2) or type-(2,0) tensors since usual linear algebra doesn't allow for a row vector of row vectors. My hunch is that in this case, transpose is used a kludge to allow expressions like   to be represented with matrices as  . Does that sound right, or am I jumping to conclusions? If this is right, it could do with some explanation somewhere. —Ben FrantzDale (talk) 15:13, 23 November 2009 (UTC)Reply

Using an orthonormal basis,  , and   That "usual linear algebra doesn't allow for a row vector of row vectors" is the reason why tensor notation is used when a row vector of row vectors, such as  , is needed. Bo Jacoby (talk) 16:53, 23 November 2009 (UTC).Reply

Also note that there is no canonical isomorphism between V and V* if V is a plain real vector space of finite dimension >1, with no additional structure. What is of course canonical is the pairing VxV* → R. Fixing a base on V is the same as fixing an isomorphism with Rn, hence produces a special isomorphism V→V*, because Rn does possess a preferred isomorphism with its dual, that is the transpose, if we represent n-vectors with columns and forms with rows. Fixing an isomorphism V→V* is the same as giving V a scalar product (check the equivalence), which is a tensor of type (0,2), that eats pairs of vectors and defecates scalars. --pma (talk) 18:46, 23 November 2009 (UTC)Reply
Those are great answers! That clarifies some things that have been nagging me for a long time! I feel like It is particularly helpful to think that conventional matrix notation doesn't provide notation for a row of row vectors or the like. I will probably copy the above discussion to Talk:Transpose for postarity and will probably add explanation along these lines to appropriate articles.
I haven't worked much with complex tensors, but your use of conjugate transpose reminds me that I've also long been suspiscious of its "meaning" (and simply that of complex conjugate) for the same reasons. Could you comment on that? In some sense   on a complex number,   is the same operation as   on a vector, using conjugate transpose as a mechanism to compute  . For a complex number, I'm not sure what would generalize to "row vector" or "column vector"... I'm not sure what I'm asking, but I feel like there's a little more that could be said connecting the above great explanations to conjugate transpose. :-) —Ben FrantzDale (talk) 19:19, 23 November 2009 (UTC)Reply
A complex number (just as a real number) is a 1-D vector, so rows and columns are the same thing. The modulus on   can be thought of as a special case of the norm on   (ie. for n=1). From an algebraic point of view, complex conjugation is the unique (non-trivial) automorphism on   that keeps   fixed. Such automorphisms are central to Galois theory. I'm not really sure what the importance and meaning is from a geometrical or analytic point of view... --Tango (talk) 19:41, 23 November 2009 (UTC)Reply

Let V and W be two vector spaces, and ƒ : VW be a linear map. Let F be the matrix representation of ƒ with respect to some bases {vi} and {wj}. I seem to recall, please do correct me if I'm wrong, that F : VW and FT : W* → V* where V* and W* are the dual spaces of V and W respectively. In this setting vT is dual to v. So the quantity vTv is the evaluation of the vector v by the covector vT. ~~ Dr Dec (Talk) ~~ 23:26, 23 November 2009 (UTC)Reply

Orthogonal Matrices edit

In the "Special transpose matrices" section, the writing implies that an orthogonal matrix G is defined as one for which GT=G-1. Thus I was going to change the "if" in "...that is, G is orthogonal if..." to "iff" but I was unsure if this was really a fundamental definition. The "Orthogonal Matrix" page does the same as this one.

It seems like a decent definition of an orthogonal matrix could be a G such that GGT and GTG are (one or both) diagonal or something. Not necessarily that one, but it's enough to make me suspect there's a more general definition some people might use.

Hopefully someone better versed in (multi-)linear algebra literature comes along and knows if there's a more general definition. If there isn't, or if it's still fully compatible, let's change the "if" to "iff" here and possibly in the "Orthogonal Matrix" page too. --Horn.imh (talk) 19:20, 16 June 2011 (UTC)Reply

I'm pretty sure you are right and that it is iff. Suppose G is not orthogonal. Then two columns of G aren't orthogonal (or a column doesn't have norm of one). Then   will not be diagonal (because off-diagonal terms are inner products of columns with different columns) in the case that the columns aren't orthogonal, and it will have something other than one on the diagonal in the case that any columns don't have norm of one (because the diagonals are the inner products of columns with themselves). QED. —Ben FrantzDale (talk) 20:05, 16 June 2011 (UTC)Reply


Notation in 'Transpose of Linear Maps' edit

  is a terrible notation for anything because it looks like a zero. We should change this to   or something. Is the author trying to suggest an  ? Because then they should just use the  ...

Also, we need to be consistent for our transpose notation. Should it be       or  ? We use three of the four possibilities here.

129.32.11.206 (talk) 19:16, 10 October 2012 (UTC)Reply

Transpose of linear maps: why defined in terms of a bilinear form? edit

In the section Transpose of linear maps, the abstract definition of a transpose is in principle independent of any bilinear form. This was stated in this way until changed by this edit (which may have been taken from Linear Algebra Quick Study Guide for Smartphones and Mobile Devices). This fundamentally changes the definition of a transpose in the abstract context. It would make more sense to me if it were defined primarily in the metric-free context, and (if desired) related to the concept defined in the section at present when suitable bilinear forms are available. I suggest reverting this section to the earlier form, with the approach using bilinear forms omitted. Does anyone with more familiarity of the area know what the most generally accepted definition is? — Quondum 14:02, 1 June 2013 (UTC)Reply

I would call what is described in that section of the article the adjoint rather than the transpose, although I'm not sure whether there is a universally accepted definition. It would make sense to me to define the transpose in a metric free setting and define the adjoint as a generalization. I'm a little surprised that we don't already have an article on the adjoint (except for the special case hermitian adjoint.) Sławomir Biały (talk) 14:59, 1 June 2013 (UTC)Reply
Thanks. It is a pity that the definitions seem to be a little variable (gauging from the few references I've browsed). I'll make a change along these lines in the next week or so, any comment from other editors being welcome. — Quondum 11:07, 2 June 2013 (UTC)Reply
I've made some comprehensive changes to the section, criticism welcome. I also removed a misguided association of the transpose of a coordinate vector and the more abstract concept of a transpose from Dual basis. — Quondum 02:00, 5 June 2013 (UTC)Reply

Error in definition of Hermitian matrix? edit

States   but I think it should be  . I think this is also what user Moala pointed out below way back in 2005. — Preceding unsigned comment added by 2601:9:2C80:464:809A:FAF:9F56:4439 (talk) 03:42, 19 May 2014 (UTC)Reply

I think this is a matter of notation: sometimes the star is used to denote complex conjugate. I've changed this in the interest of reducing ambiguity and misunderstanding. —Quondum 04:41, 19 May 2014 (UTC)Reply

Which Prime character? edit

> (also written A′, Atr, tA or At)

Should that be ′ or ʹ? Source? JDAWiseman (talk) 08:18, 9 November 2017 (UTC)Reply

Yodalearning edit

The following external link was removed:

Clicking on this link calls for enrollment in a course. If someone enrolls and finds good information on Transpose, then it might be used. For now it is just linkspam. — Rgdboer (talk) 22:59, 24 July 2018 (UTC)Reply

Matrix to the Ath power ...? edit

In the text, should "a matrix raised to the Ath power" read instead "a matrix raised to the T-th power"? — Preceding unsigned comment added by Lehnekbn (talkcontribs) 21:42, 28 January 2019 (UTC)Reply

Adjugate edit

The following was removed:

The reason transpose of a matrix is used is to get the sin(θ) between two vectors in a matrix. The dot product gives the cos(θ) of two vectors, and if we want to get the sin(θ), we would have to do a cos inverse operation to get the angle or use sin^2 + cos^2 = 1. Transpose of a matrix rotates the matrix angle to its complementary angle of pi/2 changing cos(θ) to sin(θ), allowing us to use simpler equations especially in inverse(A) = transpose(adjoint(A)) / det(A). An equation with sin(θ) and cos(θ) is vastly simpler than one with just cos(θ) to compute all other mathematical equations using tan(θ), tanh(θ), etc.

Perhaps the contributor means adjugate matrix. The business of sin(θ) being found this way is unclear and is given with no reference. — Rgdboer (talk) 22:39, 22 February 2019 (UTC)Reply

New WP:User Devssh has taken an interest in this article, made the above contribution, thanked me for the correction, and today entered more unhelpful edits into the article. Only contributes here and has not taken up a user page. Devssh is encouraged to communicate in this Talk space before contributing further. — Rgdboer (talk) 01:01, 7 March 2019 (UTC)Reply

Pertransposed edit

In some articles about numerical methods for control systems there is also the notion of the pertransposed matrix.

An example for such an article is: Varga, A. (January 1996). "Computation of Kronecker-like forms of a system pencil: applications, algorithms and software". Proceedings of Joint Conference on Control Applications Intelligent Control and Computer Aided Control System Design: 77–82. doi:10.1109/CACSD.1996.555201..

The definition there is: "transposed with respect to the main antidiagonal". (Just search for pertransposed within that article.)

If I take this literally for a square matrix   I get the pertransposed as   with Octave notation.

This is also the way it is defined in Teruel, Ginés R Pérez (2020). "Matrix operators and the Klein four group". Palestine Journal of Mathematics. 9 (1)..

But, in Voigt, Matthias (June 2010). L∞-Norm Computation for Descriptor Systems (master). Retrieved 2020-01-08. it is used as   without the transposition. Maybe that is just a typo. But, it caused me to search for pertranspose here on Wikipedia.

I needed the pertransposed in the context of the generation of the Quasi-Kronecker form of a matrix pencil (see e.g., Varga's article).

Wouldn't it be great to mention the pertranspose with the definition from articles as one of the generalizations of the transposed matrix here in this article? --TN (talk) 08:14, 8 January 2021 (UTC)Reply

Serious omission edit

For the transpose Mt of a real square matrix considered as a linear mapping M : Rn —> Rn, with the standard dot product ⟨v,w⟩, we have the standard fact that

  ⟨Mv,w⟩ = ⟨v,Mtw⟩ 

for all vectors v, w ∊ Rn.

Yet in the article this fact is buried deep in the article, in the section Adjoint, only in the greatest generality.

The section Adjoint is entirely appropriate. But the basic fact above, in its most common manifestation, ought to be mentioned much, much earlier in the article. Especially because that is often how the transpose is defined. 2601:200:C000:1A0:291B:4FAF:4C47:67FE (talk) 18:18, 24 September 2021 (UTC)Reply

Overwhelming details without the right intuition edit

This Wikipedia article can easily overwhelm a beginner. there is no point in giving all the unstructured & unstitched details. The article requires a section on motivation for transposes and the link to least squares and the four fundamental sub spaces. Can we work towards a well directed article? With a couple of volunteers I can take up this responsibility. 103.118.50.5 (talk) 05:46, 18 September 2022 (UTC)Reply

Hello Rupnagar. Using this Talk space to start, please explain the 4 fundamental subspaces. And what is the connection to least squares? Note that transpose refers to binary relations as well as linear transformations, so various details are required. As for motivation, the relation (mathematics) context is common. Rgdboer (talk) 04:02, 19 September 2022 (UTC)Reply