In mathematics, linear combinations are a concept central to linear algebra and related fields of mathematics. Most of this article deals with linear combinations in the context of a vector space over a field, with some generalisations given at the end of the article.

Definition   Suppose that K is a field and V is a vector space over K. As usual, we call elements of V vectorss and call elements of K scalars. If v1,...,vn are vectors and a1,...,an are scalars, then the linear combination of those vectors with those scalars as coefficients is
a1v1 + a2v2 + a3v3 + ... + anvn.

In a given situation, K and V may be specified explicitly, or they may be obvious from context. In that case, we often speak of a linear combination of the vectors v1,...,vn, with the coefficients unspecified (except that they must belong to K). Or, if S is a subset of V, we may speak of a linear combination of vectors in S, where both the coefficients and the vectors are unspecified, except that the vectors must belong to the set S (and the coefficients must belong to K). Finally, we may speak simply of a linear combination, where nothing is specified (except that the vectors must belong to V and the coefficients must belong to K).

Note that you can only take a linear combination of finitely many vectors (except as described in Generalisations below); that is, the number n may be finite. However, the set S that the vectors are taken from (if one is mentioned) can still be infinite; each individual linear combination will only involve finitely many vectors. Also, there's no reason that the finite number n can't be zero; in that case, we declare by convention that the result of the linear combination is the zero vector in V.

Table of contents
1 Examples and counterexamples
2 The linear span
3 Other related concepts
4 Generalisations

Examples and counterexamples

1: Examples related to analytic geometry

Let the field K be the set R of real numbers, and let the vector space V be the Euclidean space R3. Consider the vectors e1 := (1,0,0), e2 := (0,1,0) and e3 = (0,0,1). Then any vector in R3 is a linear combination of e1, e2 and e3.

To see that this is so, take an arbitrary vector (a1,a2,a3) in R3, and write:

(a1,a2,a3) = (a1,0,0) + (0,a2,0) + (0,0,a3) = a1(1,0,0) + a2(0,1,0) + a3(0,0,1) = a1e1 + a2e2 + a3e3.

2: Examples related to functional analysis

Let K be the set C of all complex numbers, and let V be the set CC(R) of all continuous functions from the real line R to the complex plane C. Consider the vectors (functions) f and g defined by f(t) := eit and g(t) := eit. (Here, e is the base of the natural logarithm, about 2.71828..., and i is the imaginary unit, a square root of −1.) Some linear combinations of f and g are:

  • cos t = (1/2)eit + (1/2)eit;
  • 2 sin t = (−i)eit + (i)eit.
On the other hand, the constant function 3 is not a linear combination of f and g. To see this, suppose that 3 could be written as a linear combination of eit and eit. This means that there would exist complex scalars a and b such that aeit + beit = 3 for all real numbers t. Setting t = 0 and t = π gives the equations a + b = 3 and a + b = −3, and clearly this cannot happen.

3: Examples related to algebraic geometry

Let K be any field (R, C, or whatever you like best), and let V be the set P of all polynomials with coefficients taken from the field K. Consider the vectors (polynomials) p1 := 1, p2 := x + 1, and p3 := x2 + x + 1.

Is the polynomial x2 − 1 a linear combination of p1, p2, and p3? To find out, consider an arbitrary linear combination of these vectors and try to see when it equals the desired vector x2 − 1. Picking arbitrary coefficients a1, a2, and a3, we want

a1(1) + a2(x + 1) + a3(x2 + x + 1) = x2 − 1.
Multiplying the polynomials out, this means
(a1) + (a2x + a2) + (a3x2 + a3x + a3) = x2 − 1,
and collecting like powers of x, we get
a3x2 + (a2 + a3)x + (a1 + a2 + a3) = 1x2 + 0x + (−1).
Two polynomials are equal
if and only if their corresponding coefficients are equal, so we can conclude
a3 = 1,    a2 + a3 = 0,    a1 + a2 + a3 = −1.
This system of linear equations can easily be solved. First, the first equation simply says that a3 is 1. Knowing that, we can solve the second equation for a2, which comes out to −1. Finally, the last equation tells us that a1 is also −1. Therefore, the only possible way to get a linear combination is with these coefficients. Indeed,
x2 − 1 = −1 − (x + 1) + (x2 + x + 1) = −p1 − p2 + p3,
so x2 − 1 is a linear combination of p1, p2, and p3.

On the other hand, what about the polynomial x3 − 1? If we try to make this vector a linear combination of p1, p2, and p3, then following the same process as before, we'll get the equation

0x3 + a3x2 + (a2 + a3)x + (a1 + a2 + a3) = 1x3 + 0x2 + 0x + (−1).
However, when we set corresponding coefficients equal in this case, the equation for x3 is
0 = 1,
which is always false. Therefore, there is no way for this to work, and x3 − 1 is not a linear combination of p1, p2, and p3.

The linear span

Take an arbitrary field K, an arbitrary vector space V, and let v1,...,vn be vectors (in V). It is interesting to consider the set of all linear combinations of these vectors. This set is called the linear span (or just span) of the vectors, say S ={v1,...,vn}. We write the span of S as span(S) or sp(S):

Sp(v1,...,vn) := {a1v1 + ... + anvn : a1,...,an ∈ K}.
(This expression is an example of set builder notation.)

Theorem 1: Sp(v1,...,vn) is a subspace of V. Furthermore, this span is the smallest subspace of V that the vectors v1,...,vn all belong to.

This fact (which is proved later in this section) is one reason why the span is important.

Now let S be a subset of the vector space V. The linear span of S consists of all linear combinations of elements of S. In symbols,

Sp(S) = {a1v1 + ... + akvk : k ∈ N, a1,...,ak ∈ K, v1,...,vk ∈ S},
where N is the set of natural numbers (including zero). Notice that this time the number of vectors involved in the linear combination can vary, from zero on up, but it must still be finite each time.

Theorem 2: Sp(S) is also a subspace of V. Furthermore, this span is the smallest subspace of V that is a superset of S.

The rest of this section is a proof of Theorem 1. Theorem 2 is very similar, but it's a bit messier to write down, since the vectors involved in any given linear combination can vary.

Proof of Theorem 1:

Property 1:
The most general possible two elements of the span are x := a1v1 + ... + anvn and y := b1v1 + ... + bnvn. We have to show that x + y is also a linear combination. By using associativity and commutativity of addition and the distributive law, we can write

x + y = (a1 + b1)v1 + ... + (an + bn)vn,
and since ai + bi is a scalar for every i, we see that x + y is indeed a linear combination of the given vectors.

Property 2:
Let c be a scalar and again take x := a1v1 + ... + anvn. We have to show that cx is also a linear combination. Now,

cx = (ca1)v1 + ... + (can)vn,
and since cai is a scalar for every i, we are done.

Property 3:
The zero element 0V of V is a linear combination because we can write

0V = 0Kv1 + 0Kv2 + ... + 0Kvn.
(Here, 0K is the zero element of the field K.) This equation is true because in every vector space we have 0Kv = 0V.

Suppose W is another subspace of V which contains the vectors v1,...,vn. Then W is closed under scalar multiplication and addition of vectors, so we can prove by mathematical induction that a1v1 + ... + anvn is an element of W for any scalars a1,...,an. Thus, sp(v1,...,vn), the set of all such linear combinations, is a subset of W.

Other related concepts

Sometimes, some single vector can be written in two different ways as a linear combination of v1,...,vn. If that is possible, then v1,...,vn are called linearly dependent; otherwise, they are linearly independent. Similarly, we can speak of linear dependence or independence of an arbitrary set S of vectors.

If S is linearly independent and the span of S equals V, then S is a basis for V.

We can think of linear combinations as the most general sort of operation on a vector space. The basic operations of addition and scalar multiplication, together with the existence of an additive identity and additive inverses, can't be combined in any more complicated way than the generic linear combination. Ultimately, this fact lies at the heart of the usefulness of linear combinations in the study of vector spaces.


If V is a topological vector space, then there may be a way to make sense of certain infinite linear combination, using the topology of V. For example, we might be able to speak of a1v1 + a2v2 + a3v3 + ..., going on forever. Such infinite linear combinations don't always make sense; we call them convergent when they do. Allowing more linear combinations in this case can also lead to a different concept of span, linear independence, and basis. The articles on the various flavours of topological vector spaces go into more detail about these.

If K is a commutative ring instead of a field, then everything that has been said above about linear combinations generalises to this case without change. The only difference is that we call spaces like V modules instead of vector spaces. If K is a noncommutative ring, then the concept still generalises, with one caveat: Since modules over noncommutative rings come in left and right versions, our linear combinations may also come in either of these versions, whatever is appropriate for the given module. This is simply a matter of doing scalar multiplication on the correct side.

A more complicated twist comes when V is a bimodule over two rings, KL and KR. In that case, the most general linear combination looks like

a1v1b1 + ... + anvnbn,
where a1,...,an belong to KL, b1,...,bn belong to KR, and v1,...,vn belong to V.