Skip to content
Surf Wiki
Save to docs
science/mathematics

From Surf Wiki (app.surf) — the open knowledge base

Basis (linear algebra)

Set of vectors used to define coordinates

Basis (linear algebra)

Set of vectors used to define coordinates

The same vector (dark purple) can be represented in two different bases (purple and red arrows).

In mathematics, a set B of elements of a vector space V is called a basis (: bases) if every element of V can be written in a unique way as a finite linear combination of elements of B. The coefficients of this linear combination are referred to as components or coordinates of the vector with respect to B. The elements of a basis are called basis vectors.

Equivalently, a set B is a basis if its elements are linearly independent and every element of V is a linear combination of elements of B. In other words, a basis is a linearly independent spanning set.

A vector space can have several bases; however all the bases have the same number of elements, called the dimension of the vector space.

This article deals mainly with finite-dimensional vector spaces. However, many of the principles are also valid for infinite-dimensional vector spaces.

Basis vectors find applications in the study of crystal structures and frames of reference.

Definition

A basis B of a vector space V over a field F (such as the real numbers R or the complex numbers C) is a linearly independent subset of V that spans V. This means that a subset B of V is a basis if it satisfies the two following conditions:

  • linear independence: for every finite subset {\mathbf v_1, \dotsc, \mathbf v_m} of B, if c_1 \mathbf v_1 + \cdots + c_m \mathbf v_m = \mathbf 0 for some c_1,\dotsc,c_m in F, then c_1 = \cdots = c_m = 0; and
  • the spanning property: for every vector \mathbf{v} in V , one can choose a_1,\dotsc,a_n in F and \mathbf v_1, \dotsc, \mathbf v_n in B such that In other words, \mathbb{v} can be represented as a linear combination of some vectors in B .

The first property may be equivalently phrased as follows: if a linear combination of vectors in B is equal to the zero vector, then all the scalars coefficients of the combination must be zero.

If B is a basis for V, then every vector \mathbf{v} in V can be written as a linear combination of vectors in B (by the spanning property); it follows from linear independence that this can be done in exactly one way. Thus, the scalar coefficients a_i that appear in this combination are uniquely determined; they are called the coordinates of \mathbf{v} with respect to the basis B .

A vector space that has a finite basis is called finite-dimensional. In this case, the finite subset can be taken as B itself to check for linear independence in the above definition.

It is often convenient or even necessary to have an ordering on the basis vectors, for example, when discussing orientation, or when one considers the scalar coefficients of a vector with respect to a basis without referring explicitly to the basis elements. In this case, the ordering is necessary for associating each coefficient with the corresponding basis element. This ordering can be done by numbering the basis elements. In order to emphasize that an order has been chosen, one speaks of an ordered basis, which is therefore not simply an unstructured set, but a sequence, an indexed family, or similar; see below.

Examples

The set R2 of the ordered pairs of real numbers is a vector space under the operations of component-wise addition (a, b) + (c, d) = (a + c, b+d) and scalar multiplication \lambda (a,b) = (\lambda a, \lambda b), where \lambda is any real number. A simple basis of this vector space consists of the two vectors and . These vectors form a basis (called the standard basis) because any vector of R2 may be uniquely written as \mathbf v = a \mathbf e_1 + b \mathbf e_2. Any other pair of linearly independent vectors of R2, such as (1, 1) and (−1, 2), forms also a basis of R2.

More generally, if F is a field, the set F^n of n-tuples of elements of F is a vector space for similarly defined addition and scalar multiplication. Let \mathbf e_i = (0, \ldots, 0,1,0,\ldots, 0) be the n-tuple with all components equal to 0, except the ith, which is 1. Then \mathbf e_1, \ldots, \mathbf e_n is a basis of F^n, which is called the standard basis of F^n.

A different flavor of example is given by polynomial rings. If F is a field, the collection F[X] of all polynomials in one indeterminate X with coefficients in F is an F-vector space. One basis for this space is the monomial basis B, consisting of all monomials: B={1, X, X^2, \ldots}. Any set of polynomials such that there is exactly one polynomial of each degree (such as the Bernstein basis polynomials or Chebyshev polynomials) is also a basis. (Such a set of polynomials is called a polynomial sequence.) But there are also many bases for F[X] that are not of this form.

Properties

Many properties of finite bases result from the Steinitz exchange lemma, which states that, for any vector space V, given a finite spanning set S and a linearly independent set L of n elements of V, one may replace n well-chosen elements of S by the elements of L to get a spanning set containing L, having its other elements in S, and having the same number of elements as S.

Most properties resulting from the Steinitz exchange lemma remain true when there is no finite spanning set, but their proofs in the infinite case generally require the axiom of choice or a weaker form of it, such as the ultrafilter lemma.

If V is a vector space over a field F, then:

  • If L is a linearly independent subset of a spanning set SV, then there is a basis B such that L\subseteq B\subseteq S.
  • V has a basis (this is the preceding property with L being the empty set, and ).
  • All bases of V have the same cardinality, which is called the dimension of V. This is the dimension theorem.
  • A generating set S is a basis of V if and only if it is minimal, that is, no proper subset of S is also a generating set of V.
  • A linearly independent set L is a basis if and only if it is maximal, that is, it is not a proper subset of any linearly independent set.

If V is a vector space of dimension n, then:

  • A subset of V with n elements is a basis if and only if it is linearly independent.
  • A subset of V with n elements is a basis if and only if it is a spanning set of V.

Coordinates {{anchor|Ordered bases and coordinates}}

Let V be a vector space of finite dimension n over a field F, and B = {\mathbf b_1, \ldots, \mathbf b_n} be a basis of V. By definition of a basis, every v in V may be written, in a unique way, as \mathbf v = \lambda_1 \mathbf b_1 + \cdots + \lambda_n \mathbf b_n, where the coefficients \lambda_1, \ldots, \lambda_n are scalars (that is, elements of F), which are called the coordinates of v over B. However, if one talks of the set of the coefficients, one loses the correspondence between coefficients and basis elements, and several vectors may have the same set of coefficients. For example, 3 \mathbf b_1 + 2 \mathbf b_2 and 2 \mathbf b_1 + 3 \mathbf b_2 have the same set of coefficients {2, 3}, and are different. It is therefore often convenient to work with an ordered basis; this is typically done by indexing the basis elements by the first natural numbers. Then, the coordinates of a vector form a sequence similarly indexed, and a vector is completely characterized by the sequence of coordinates. An ordered basis, especially when used in conjunction with an origin, is also called a coordinate frame or simply a frame (for example, a Cartesian frame or an affine frame).

Let, as usual, F^n be the set of the n-tuples of elements of F. This set is an F-vector space, with addition and scalar multiplication defined component-wise. The map \varphi: (\lambda_1, \ldots, \lambda_n) \mapsto \lambda_1 \mathbf b_1 + \cdots + \lambda_n \mathbf b_n is a linear isomorphism from the vector space F^n onto V. In other words, F^n is the coordinate space of V, and the n-tuple \varphi^{-1}(\mathbf v) is the coordinate vector of v.

The inverse image by \varphi of \mathbf b_i is the n-tuple \mathbf e_i all of whose components are 0, except the ith that is 1. The \mathbf e_i form an ordered basis of F^n, which is called its standard basis or canonical basis. The ordered basis B is the image by \varphi of the canonical basis of F^n.

It follows from what precedes that every ordered basis is the image by a linear isomorphism of the canonical basis of F^n, and that every linear isomorphism from F^n onto V may be defined as the isomorphism that maps the canonical basis of F^n onto a given ordered basis of V. In other words, it is equivalent to define an ordered basis of V, or a linear isomorphism from F^n onto V.

Change of basis

Main article: Change of basis

Let V be a vector space of dimension n over a field F. Given two (ordered) bases B_\text{old} = (\mathbf v_1, \ldots, \mathbf v_n) and B_\text{new} = (\mathbf w_1, \ldots, \mathbf w_n) of V, it is often useful to express the coordinates of a vector x with respect to B_\mathrm{old} in terms of the coordinates with respect to B_\mathrm{new}. This can be done by the change-of-basis formula, that is described below. The subscripts "old" and "new" have been chosen because it is customary to refer to B_\mathrm{old} and B_\mathrm{new} as the old basis and the new basis, respectively. It is useful to describe the old coordinates in terms of the new ones, because, in general, one has expressions involving the old coordinates, and if one wants to obtain equivalent expressions in terms of the new coordinates; this is obtained by replacing the old coordinates by their expressions in terms of the new coordinates.

Typically, the new basis vectors are given by their coordinates over the old basis, that is, \mathbf w_j = \sum_{i=1}^n a_{i,j} \mathbf v_i. If (x_1, \ldots, x_n) and (y_1, \ldots, y_n) are the coordinates of a vector x over the old and the new basis respectively, the change-of-basis formula is x_i = \sum_{j=1}^n a_{i,j}y_j, for .

This formula may be concisely written in matrix notation. Let A be the matrix of the a_{i,j}, and X= \begin{bmatrix} x_1 \ \vdots \ x_n \end{bmatrix} \quad \text{and} \quad Y = \begin{bmatrix} y_1 \ \vdots \ y_n \end{bmatrix} be the column vectors of the coordinates of v in the old and the new basis respectively, then the formula for changing coordinates is X = A Y.

The formula can be proven by considering the decomposition of the vector x on the two bases: one has \mathbf x = \sum_{i=1}^n x_i \mathbf v_i, and \mathbf x =\sum_{j=1}^n y_j \mathbf w_j = \sum_{j=1}^n y_j\sum_{i=1}^n a_{i,j}\mathbf v_i = \sum_{i=1}^n \biggl(\sum_{j=1}^n a_{i,j}y_j\biggr)\mathbf v_i.

The change-of-basis formula results then from the uniqueness of the decomposition of a vector over a basis, here B_\text{old}; that is x_i = \sum_{j=1}^n a_{i,j} y_j, for .

Proof that every vector space has a basis

Let V be any vector space over some field F. Let X be the set of all linearly independent subsets of V.

The set X is nonempty since the empty set is an independent subset of V, and it is partially ordered by inclusion, which is denoted, as usual, by ⊆.

Let Y be a subset of X that is totally ordered by ⊆, and let LY be the union of all the elements of Y (which are themselves certain subsets of V).

Since (Y, ⊆) is totally ordered, every finite subset of LY is a subset of an element of Y, which is a linearly independent subset of V, and hence LY is linearly independent. Thus LY is an element of X. Therefore, LY is an upper bound for Y in (X, ⊆): it is an element of X, that contains every element of Y.

As X is nonempty, and every totally ordered subset of (X, ⊆) has an upper bound in X, Zorn's lemma asserts that X has a maximal element. In other words, there exists some element Lmax of X satisfying the condition that whenever Lmax ⊆ L for some element L of X, then .

It remains to prove that Lmax is a basis of V. Since Lmax belongs to X, we already know that Lmax is a linearly independent subset of V.

If there were some vector w of V that is not in the span of Lmax, then w would not be an element of Lmax either. Let }. This set is an element of X, that is, it is a linearly independent subset of V (because w is not in the span of Lmax, and Lmax is independent). As Lmax ⊆ Lw, and Lmax ≠ Lw (because Lw contains the vector w that is not contained in Lmax), this contradicts the maximality of Lmax. Thus this shows that Lmax spans V.

Hence Lmax is linearly independent and spans V. It is thus a basis of V, and this proves that every vector space has a basis.

This proof relies on Zorn's lemma, which is equivalent to the axiom of choice. Conversely, it has been proved that if every vector space has a basis, then the axiom of choice is true. Thus the two assertions are equivalent.

Notes

References

General references

Historical references

  • , reprint:

References

  1. {{Harvnb. Hamel. 1905
  2. Note that one cannot say "most" because the cardinalities of the two sets (functions that can and cannot be represented with a finite number of basis functions) are the same.
  3. Rees, Elmer G.. (2005). "Notes on Geometry". Springer.
  4. Kuczma, Marek. (1970). "Some remarks about additive functions on cones". [[Aequationes Mathematicae]].
  5. (1995). "Stochastic choice of basis functions in adaptive function approximation and the functional-link net". IEEE Trans. Neural Netw..
  6. Artstein, Shiri. (2002). "Proportional concentration phenomena of the sphere". [[Israel Journal of Mathematics]].
  7. {{Harvnb. Blass. 1984
Info: Wikipedia Source

This article was imported from Wikipedia and is available under the Creative Commons Attribution-ShareAlike 4.0 License. Content has been adapted to SurfDoc format. Original contributors can be found on the article history page.

Want to explore this topic further?

Ask Mako anything about Basis (linear algebra) — get instant answers, deeper analysis, and related topics.

Research with Mako

Free with your Surf account

Content sourced from Wikipedia, available under CC BY-SA 4.0.

This content may have been generated or modified by AI. CloudSurf Software LLC is not responsible for the accuracy, completeness, or reliability of AI-generated content. Always verify important information from primary sources.

Report