BivectorIn mathematics, a bivector or 2-vector is a quantity in exterior algebra or geometric algebra that extends the idea of scalars and vectors. If a scalar is considered a degree-zero quantity, and a vector is a degree-one quantity, then a bivector can be thought of as being of degree two. Bivectors have applications in many areas of mathematics and physics. They are related to complex numbers in two dimensions and to both pseudovectors and quaternions in three dimensions.
DyadicsIn mathematics, specifically multilinear algebra, a dyadic or dyadic tensor is a second order tensor, written in a notation that fits in with vector algebra. There are numerous ways to multiply two Euclidean vectors. The dot product takes in two vectors and returns a scalar, while the cross product returns a pseudovector. Both of these have various significant geometric interpretations and are widely used in mathematics, physics, and engineering. The dyadic product takes in two vectors and returns a second order tensor called a dyadic in this context.
Exterior algebraIn mathematics, the exterior algebra, or Grassmann algebra, named after Hermann Grassmann, is an algebra that uses the exterior product or wedge product as its multiplication. In mathematics, the exterior product or wedge product of vectors is an algebraic construction used in geometry to study areas, volumes, and their higher-dimensional analogues. The exterior product of two vectors and , denoted by is called a bivector and lives in a space called the exterior square, a vector space that is distinct from the original space of vectors.
Multilinear mapIn linear algebra, a multilinear map is a function of several variables that is linear separately in each variable. More precisely, a multilinear map is a function where and are vector spaces (or modules over a commutative ring), with the following property: for each , if all of the variables but are held constant, then is a linear function of . A multilinear map of one variable is a linear map, and of two variables is a bilinear map. More generally, a multilinear map of k variables is called a k-linear map.
Hermann GrassmannHermann Günther Grassmann (Graßmann, ˈhɛʁman ˈɡʏntɐ ˈɡʁasman; 15 April 1809 – 26 September 1877) was a German polymath known in his day as a linguist and now also as a mathematician. He was also a physicist, general scholar, and publisher. His mathematical work was little noted until he was in his sixties. His work preceded and exceeded the concept which is now known as a vector space. He introduced the Grassmannian, the space which parameterizes all k-dimensional linear subspaces of an n-dimensional vector space V.
Symmetric tensorIn mathematics, a symmetric tensor is a tensor that is invariant under a permutation of its vector arguments: for every permutation σ of the symbols {1, 2, ..., r}. Alternatively, a symmetric tensor of order r represented in coordinates as a quantity with r indices satisfies The space of symmetric tensors of order r on a finite-dimensional vector space V is naturally isomorphic to the dual of the space of homogeneous polynomials of degree r on V.
Levi-Civita symbolIn mathematics, particularly in linear algebra, tensor analysis, and differential geometry, the Levi-Civita symbol or Levi-Civita epsilon represents a collection of numbers; defined from the sign of a permutation of the natural numbers 1, 2, ..., n, for some positive integer n. It is named after the Italian mathematician and physicist Tullio Levi-Civita. Other names include the permutation symbol, antisymmetric symbol, or alternating symbol, which refer to its antisymmetric property and definition in terms of permutations.
Tensor (intrinsic definition)In mathematics, the modern component-free approach to the theory of a tensor views a tensor as an abstract object, expressing some definite type of multilinear concept. Their properties can be derived from their definitions, as linear maps or more generally; and the rules for manipulations of tensors arise as an extension of linear algebra to multilinear algebra. In differential geometry an intrinsic geometric statement may be described by a tensor field on a manifold, and then doesn't need to make reference to coordinates at all.
MultivectorIn multilinear algebra, a multivector, sometimes called Clifford number, is an element of the exterior algebra Λ(V) of a vector space V. This algebra is graded, associative and alternating, and consists of linear combinations of simple k-vectors (also known as decomposable k-vectors or k-blades) of the form where are in V. A k-vector is such a linear combination that is homogeneous of degree k (all terms are k-blades for the same k).
Tensor calculusIn mathematics, tensor calculus, tensor analysis, or Ricci calculus is an extension of vector calculus to tensor fields (tensors that may vary over a manifold, e.g. in spacetime). Developed by Gregorio Ricci-Curbastro and his student Tullio Levi-Civita, it was used by Albert Einstein to develop his general theory of relativity. Unlike the infinitesimal calculus, tensor calculus allows presentation of physics equations in a form that is independent of the choice of coordinates on the manifold.
Minor (linear algebra)In linear algebra, a minor of a matrix A is the determinant of some smaller square matrix, cut down from A by removing one or more of its rows and columns. Minors obtained by removing just one row and one column from square matrices (first minors) are required for calculating matrix cofactors, which in turn are useful for computing both the determinant and inverse of square matrices. The requirement that the square matrix be smaller than the original matrix is often omitted in the definition.
Tensor algebraIn mathematics, the tensor algebra of a vector space V, denoted T(V) or T^•(V), is the algebra of tensors on V (of any rank) with multiplication being the tensor product. It is the free algebra on V, in the sense of being left adjoint to the forgetful functor from algebras to vector spaces: it is the "most general" algebra containing V, in the sense of the corresponding universal property (see below). The tensor algebra is important because many other algebras arise as quotient algebras of T(V).
Rotation (mathematics)Rotation in mathematics is a concept originating in geometry. Any rotation is a motion of a certain space that preserves at least one point. It can describe, for example, the motion of a rigid body around a fixed point. Rotation can have a sign (as in the sign of an angle): a clockwise rotation is a negative magnitude so a counterclockwise turn has a positive magnitude. A rotation is different from other types of motions: translations, which have no fixed points, and (hyperplane) reflections, each of them having an entire (n − 1)-dimensional flat of fixed points in a n-dimensional space.
Multilinear formIn abstract algebra and multilinear algebra, a multilinear form on a vector space over a field is a map that is separately -linear in each of its arguments. More generally, one can define multilinear forms on a module over a commutative ring. The rest of this article, however, will only consider multilinear forms on finite-dimensional vector spaces. A multilinear -form on over is called a (covariant) -tensor, and the vector space of such forms is usually denoted or .
PseudovectorIn physics and mathematics, a pseudovector (or axial vector) is a quantity that behaves like a vector in many situations, but its direction does not conform when the object is rigidly transformed by rotation, translation, reflection, etc. This can also happen when the orientation of the space is changed. For example, the angular momentum is a pseudovector because it is often described as a vector, but by just changing the position of reference (and changing the position vector), angular momentum can reverse direction, which is not supposed to happen with true vectors.
TensorIn mathematics, a tensor is an algebraic object that describes a multilinear relationship between sets of algebraic objects related to a vector space. Tensors may map between different objects such as vectors, scalars, and even other tensors. There are many types of tensors, including scalars and vectors (which are the simplest tensors), dual vectors, multilinear maps between vector spaces, and even some operations such as the dot product.
AlgebraAlgebra () is the study of variables and the rules for manipulating these variables in formulas; it is a unifying thread of almost all of mathematics. Elementary algebra deals with the manipulation of variables (commonly represented by Roman letters) as if they were numbers and is therefore essential in all applications of mathematics. Abstract algebra is the name given, mostly in education, to the study of algebraic structures such as groups, rings, and fields.
Outer productIn linear algebra, the outer product of two coordinate vectors is the matrix whose entries are all products of an element in the first vector with an element in the second vector. If the two coordinate vectors have dimensions n and m, then their outer product is an n × m matrix. More generally, given two tensors (multidimensional arrays of numbers), their outer product is a tensor. The outer product of tensors is also referred to as their tensor product, and can be used to define the tensor algebra.
PseudoscalarIn linear algebra, a pseudoscalar is a quantity that behaves like a scalar, except that it changes sign under a parity inversion while a true scalar does not. Any scalar product between a pseudovector and an ordinary vector is a pseudoscalar. The prototypical example of a pseudoscalar is the scalar triple product, which can be written as the scalar product between one of the vectors in the triple product and the cross product between the two other vectors, where the latter is a pseudovector.
Covariance and contravariance of vectorsIn physics, especially in multilinear algebra and tensor analysis, covariance and contravariance describe how the quantitative description of certain geometric or physical entities changes with a change of basis. In modern mathematical notation, the role is sometimes swapped. A simple illustrative case is that of a vector. For a vector, once a set of basis vectors has been defined, then the components of that vector will always vary opposite to that of the basis vectors. A vector is therefore a contravariant tensor.