# Tensor algebra

In mathematics, the **tensor algebra** of a vector space *V*, denoted *T*(*V*) or *T*^{ •}(*V*), is the algebra of tensors on *V* (of any rank) with multiplication being the tensor product. It is the free algebra on *V*, in the sense of being left adjoint to the forgetful functor from algebras to vector spaces: it is the "most general" algebra containing *V*, in the sense of the corresponding universal property (see below).

The tensor algebra is important because many other algebras arise as quotient algebras of *T*(*V*). These include the exterior algebra, the symmetric algebra, Clifford algebras, the Weyl algebra and universal enveloping algebras.

The tensor algebra also has two coalgebra structures; one simple one, which does not make it a bialgebra, but does lead to the concept of a cofree coalgebra, and a more complicated one, which yields a bialgebra, and can be extended by giving an antipode to create a Hopf algebra structure.

*Note*: In this article, all algebras are assumed to be unital and associative. The unit is explicitly required to define the coproduct.

## Construction

Let *V* be a vector space over a field *K*. For any nonnegative integer *k*, we define the ** kth tensor power** of

*V*to be the tensor product of

*V*with itself

*k*times:

That is, *T*^{k}*V* consists of all tensors on *V* of order *k*. By convention *T*^{0}*V* is the ground field *K* (as a one-dimensional vector space over itself).

We then construct *T*(*V*) as the direct sum of *T*^{k}*V* for *k* = 0,1,2,…

The multiplication in *T*(*V*) is determined by the canonical isomorphism

given by the tensor product, which is then extended by linearity to all of *T*(*V*). This multiplication rule implies that the tensor algebra *T*(*V*) is naturally a graded algebra with *T*^{k}*V* serving as the grade-*k* subspace. This grading can be extended to a **Z** grading by appending subspaces for negative integers *k*.

The construction generalizes in straightforward manner to the tensor algebra of any module *M* over a *commutative* ring. If *R* is a non-commutative ring, one can still perform the construction for any *R*-*R* bimodule *M*. (It does not work for ordinary *R*-modules because the iterated tensor products cannot be formed.)

## Adjunction and universal property

The tensor algebra *T*(*V*) is also called the **free algebra** on the vector space *V*, and is functorial. As with other free constructions, the functor *T* is left adjoint to some forgetful functor. In this case, it's the functor which sends each *K*-algebra to its underlying vector space.

Explicitly, the tensor algebra satisfies the following universal property, which formally expresses the statement that it is the most general algebra containing *V*:

- Any linear transformation
*f*:*V*→*A*from*V*to an algebra*A*over*K*can be uniquely extended to an algebra homomorphism from*T*(*V*) to*A*as indicated by the following commutative diagram:

Here *i* is the canonical inclusion of *V* into *T*(*V*) (the unit of the adjunction). One can, in fact, define the tensor algebra *T*(*V*) as the unique algebra satisfying this property (specifically, it is unique up to a unique isomorphism), but one must still prove that an object satisfying this property exists.

The above universal property shows that the construction of the tensor algebra is *functorial* in nature. That is, *T* is a functor from ** K-Vect**, the category of vector spaces over

*K*, to

**, the category of**

*K*-Alg*K*-algebras. The functoriality of

*T*means that any linear map from

*V*to

*W*extends uniquely to an algebra homomorphism from

*T*(

*V*) to

*T*(

*W*).

## Non-commutative polynomials

If *V* has finite dimension *n*, another way of looking at the tensor algebra is as the "algebra of polynomials over *K* in *n* non-commuting variables". If we take basis vectors for *V*, those become non-commuting variables (or *indeterminants*) in *T*(*V*), subject to no constraints beyond associativity, the distributive law and *K*-linearity.

Note that the algebra of polynomials on *V* is not , but rather : a (homogeneous) linear function on *V* is an element of for example coordinates on a vector space are covectors, as they take in a vector and give out a scalar (the given coordinate of the vector).

## Quotients

Because of the generality of the tensor algebra, many other algebras of interest can be constructed by starting with the tensor algebra and then imposing certain relations on the generators, i.e. by constructing certain quotient algebras of *T*(*V*). Examples of this are the exterior algebra, the symmetric algebra, Clifford algebras, the Weyl algebra and universal enveloping algebras.

## Coalgebra

The tensor algebra has two different coalgebra structures. One is compatible with the tensor product, and thus can be extended to a bialgebra, and can be further be extended with an antipode to a Hopf algebra structure. The other structure, although simpler, cannot be extended to a bialgebra. The first structure is developed immediately below; the second structure is given in the section on the cofree coalgebra, further down.

The development provided below can be equally well applied to the exterior algebra, using the wedge symbol in place of the tensor symbol ; a sign must also be kept track of, when permuting elements of the exterior algebra. This correspondence also lasts through the definition of the bialgebra, and on to the definition of a Hopf algebra. That is, the exterior algebra can also be given a Hopf algebra structure.

Similarly, the symmetric algebra can also be given the structure of a Hopf algebra, in exactly the same fashion, by replacing everywhere the tensor product by the symmetrized tensor product , i.e. that product where

In each case, this is possible because the alternating product and the symmetric product obey the required consistency conditions for the definition of a bialgebra and Hopf algebra; this can be explicitly checked in the manner below. Whenever one has a product obeying these consistency conditions, the construction goes thorough; insofar as such a product gave rise to a quotient space, the quotient space inherits the Hopf algebra structure.

In the language of category theory, one says that there is a functor *T* from the category of *K*-vector spaces to the category of *K*-associate algebras. But there is also a functor *Λ* taking vector spaces to the category of exterior algebras, and a functor *Sym* taking vector spaces to symmetric algebras. There is a natural map from *T* to each of these. Verifying that quotienting preserves the Hopf algebra structure is the same as verifying that the maps are indeed natural.

### Coproduct

The coalgebra is obtained by defining a coproduct or diagonal operator

Here, is used as a short-hand for to avoid an explosion of parenthesis. The symbol is used to denote the "external" tensor product, needed for the definition of a coalgebra. It is being used to distinguish it from the "internal" tensor product , which is already "taken" and being used to denote multiplication in the tensor algebra (see the section *Multiplication*, below, for further clarification on this issue). In order to avoid confusion between these two symbols, most texts will replace by a plain dot, or even drop it altogether, with the understanding that it is implied from context. This then allows the otimes symbol to be used in place of the boxtimes symbol. This is not done below, and the two symbols are used independently and explicitly, so as to show the proper location of each. The result is a bit more verbose, but should be easier to comprehend.

The definition of the operator is most easily built up in stages, first by defining it for elements and then by homomorphically extending it to the whole algebra. A suitable choice for the coproduct is then

and

where is the unit of the field . By linearity, one obviously has

for all It is straight-forward to verify that this definition satisfies the axioms of a coalgebra: that is, that

where is the identity map on . Indeed, one gets

and likewise for the other side. At this point, one could invoke a lemma, and say that extends trivially, by linearity, to all of , because is a free object and is a generator of the free algebra, and is a homomorphism. However, it is insightful to provide explicit expressions. So, for , one has (by definition) the homomorphism

Expanding, one has

In the above expansion, there is no need to ever write as this is just plain-old scalar multiplication in the algebra; that is, one trivially has that

The extension above preserves the algebra grading. That is,

Continuing in this fashion, one can obtain an explicit expression for the coproduct acting on a homogenous element of order *m*:

where the ш symbol, the sha, denotes the shuffle product. This is expressed in the second summation, which is taken over all (p,m-p+1)-shuffles. The above is written with a notational trick, to keep track of the field element 1: the trick is to write , and this is shuffled into various locations during the expansion of the sum over shuffles. The shuffle follows directly from the first axiom of a co-algebra: the relative order of the elements is *preserved* in the riffle shuffle: the riffle shuffle merely splits the ordered sequence into two ordered sequences, one on the left, and one on the right. Any one given shuffle obeys

As before, the algebra grading is preserved:

### Counit

The counit is the terminal object of the algebra, and is given by the projection of the field component out from the algebra. This can be written as for and for . By homomorphism under the tensor product , this extends to

for all It is a straight-forward matter to verify that this counit satisfies the needed axiom for the coalgebra:

Working this explicitly, one has

where, for the last step, one has made use of the isomorphism , as is appropriate for the defining axiom of the counit.

## Bialgebra

A bialgebra defines both multiplication, and comultiplication, and requires them to be compatible.

### Multiplication

Multiplication is given by an operator

which, in this case, was already given as the "internal" tensor product. That is,

That is, The above should make it clear why the symbol needs to be used: the was actually one and the same thing as ; and notational sloppiness here would lead to utter chaos. To strengthen this: the tensor product of the tensor algebra corresponds to the multiplication used in the definition of a (bi)-algebra, whereas the tensor product is the one required in the definition of a (bi)-algebra. These two tensor products are *not* the same thing!

### Unit

The unit is the initial object for the algebra, and is exactly as expected:

is just the embedding, so that

That the unit is compatible with the tensor product is "trivial": it is just part of the standard definition of the tensor product of vector spaces. That is, for field element *k* and any More verbosely, the axioms for an associative algebra require the two homomorphisms (or commuting diagrams):

on , and that symmetrically, on , that

where the right-hand side of these equations should be understood as the scalar product.

### Compatibility

The unit and counit, and multiplication and comultiplication, all have to satisfy compatibility conditions. It is straightforward to see that

Similarly, the unit is compatible with comultiplication:

The above requires the use of the isomorphism in order to work; without this, one looses linearity. Component-wise,

with the right-hand side making use of the isomorphism.

Multiplication and the counit are compatible:

whenever *x* or *y* are not elements of , and otherwise, one has scalar multiplication on the field: The most difficult to verify is the compatibility of multiplication and comultiplication:

where exchanges elements. The compatibility condition only needs to be verified on ; the full compatibility follows as a homomorphic extension to all of The verification is verbose but straight-forward; it is not given here, except for the final result:

For an explicit expression for this was given in the coalgebra section, above.

## Hopf algebra

The Hopf algebra adds an antipode to the bialgebra axioms. The antipode on is given by

This is sometimes called the "anti-identity". The antipode on is given by

and on by

This extends homomorphically to

### Compatibility

Compatibility of the antipode with multiplication and comultiplication requires that

This is straight-forward to verify componentwise on :

Similarly, on :

Recall that

and that

for any that is *not* in

One may proceed in a similar manner, by homomorphism, verifying that the antipode inserts the appropriate cancellative signs in the shuffle, starting with the compatibility condition on and proceeding by induction.

## Cofree coalgebra

One may define a different coproduct on the tensor algebra, simpler than the one given above. It is given by

Here, as before, one uses the notation trick that (and recalling that trivially).

This coproduct does give rise to a coalgebra, however, it *cannot* be turned into a bialgebra. It does, however, describe a coalgebra that is dual to the algebra structure on *T*(*V*^{∗}), where *V*^{∗} denotes the dual vector space of linear maps *V* → **F**. In the same way that the tensor algebra is a free algebra, the corresponding coalgebra is termed co-free.

The difference between this, and the other coalgebra is most easily seen in the term. Here, one has that

for , which is clearly missing a shuffled term, as compared to before.

## See also

- Braided vector space
- Braided Hopf algebra
- Monoidal category
- Multilinear subspace learning
- Stanisław Lem's
*Love and Tensor Algebra* - Fock space

## References

- Bourbaki, Nicolas (1989).
*Algebra I. Chapters 1-3*. Elements of Mathematics. Springer-Verlag. ISBN 3-540-64243-9.*(See Chapter 3 §5)*

- Serge Lang (2002),
*Algebra*, Graduate Texts in Mathematics,**211**(3rd ed.), Springer Verlag, ISBN 978-0-387-95385-4