In mathematics, especially in the field of algebra, a polynomial ring or polynomial algebra is a ring formed from the set of polynomials in one or more indeterminates (traditionally also called variables) with coefficients in another ring, often a field.
Often, the term "polynomial ring" refers implicitly to the special case of a polynomial ring in one indeterminate over a field. The importance of such polynomial rings relies on the high number of properties that they have in common with the ring of the integers.[1]
Polynomial rings occur and are often fundamental in many parts of mathematics such as number theory, commutative algebra, and algebraic geometry. In ring theory, many classes of rings, such as unique factorization domains, regular rings, group rings, rings of formal power series, Ore polynomials, graded rings, have been introduced for generalizing some properties of polynomial rings.[2]
A closely related notion is that of the ring of polynomial functions on a vector space, and, more generally, ring of regular functions on an algebraic variety.
Let be a field or (more generally) a commutative ring.
The polynomial ring in over, which is denoted, can be defined in several equivalent ways. One of them is to define as the set of expressions, called polynomials in, of the form
p=p0+p1X+p2X2+ … +pmXm+pmXm,
XkXl=Xk+l
Two polynomials are equal when the corresponding coefficients of each are equal.
One can think of the ring as arising from by adding one new element that is external to, commutes with all elements of, and has no other specific properties. This can be used for an equivalent definition of polynomial rings.
The polynomial ring in over is equipped with an addition, a multiplication and a scalar multiplication that make it a commutative algebra. These operations are defined according to the ordinary rules for manipulating algebraic expressions. Specifically, if
p=p0+p1X+p2X2+ … +pmXm,
and
q=q0+q1X+q2X2+ … +qnXn,
then
p+q=r0+r1X+r2X2+ … +rkXk,
and
pq=s0+s1X+s2X2+ … +slXl,
where,
ri=pi+qi
and
si=p0qi+p1qi-1+ … +piq0.
The scalar multiplication is the special case of the multiplication where is reduced to its constant term (the term that is independent of); that is
p0\left(q0+q1X+...+qnXn\right)=p0q0+\left(p0q1\right)X+ … +\left(p0
n | |
q | |
n\right)X |
It is straightforward to verify that these three operations satisfy the axioms of a commutative algebra over . Therefore, polynomial rings are also called polynomial algebras.
Another equivalent definition is often preferred, although less intuitive, because it is easier to make it completely rigorous, which consists in defining a polynomial as an infinite sequence of elements of, having the property that only a finite number of the elements are nonzero, or equivalently, a sequence for which there is some so that for . In this case, and are considered as alternate notations for the sequences and, respectively. A straightforward use of the operation rules shows that the expression
p0+p1X+p2X2+ … +pmXm
.
Let
p=p0+p1X+p2X2+ … +pmXm+pmXm,
pm\ne0
The constant term of is
p0.
The degree of, written is
m,
The leading coefficient of is
pm.
In the special case of the zero polynomial, all of whose coefficients are zero, the leading coefficient is undefined, and the degree has been variously left undefined,[4] defined to be,[5] or defined to be a .[6]
A constant polynomial is either the zero polynomial, or a polynomial of degree zero.
A nonzero polynomial is monic if its leading coefficient is
1.
Given two polynomials and, if the degree of the zero polynomial is defined to be
-infty,
\deg(p+q)\lemax(\deg(p),\deg(q)),
\deg(pq)=\deg(p)+\deg(q).
It follows immediately that, if is an integral domain, then so is .
It follows also that, if is an integral domain, a polynomial is a unit (that is, it has a multiplicative inverse) if and only if it is constant and is a unit in .
Two polynomials are associated if either one is the product of the other by a unit.
Over a field, every nonzero polynomial is associated to a unique monic polynomial.
Given two polynomials, and, one says that divides, is a divisor of, or is a multiple of, if there is a polynomial such that .
A polynomial is irreducible if it is not the product of two non-constant polynomials, or equivalently, if its divisors are either constant polynomials or have the same degree.
Let be a field or, more generally, a commutative ring, and a ring containing . For any polynomial in and any element in, the substitution of with in defines an element of, which is denoted . This element is obtained by carrying on in after the substitution the operations indicated by the expression of the polynomial. This computation is called the evaluation of at . For example, if we have
P=X2-1,
\begin{align} P(3)&=32-1=8,\\ P(X2+1)&=\left(X2+1\right)2-1=X4+2X2 \end{align}
(in the first example, and in the second one). Substituting for itself results in
P=P(X),
explaining why the sentences "Let be a polynomial" and "Let be a polynomial" are equivalent.
The polynomial function defined by a polynomial is the function from into that is defined by
x\mapstoP(x).
For every in, the evaluation at, that is, the map
P\mapstoP(a)
For every ring containing, and every element of, there is a unique algebra homomorphism from to that fixes, and maps to .
As for all universal properties, this defines the pair up to a unique isomorphism, and can therefore be taken as a definition of .
The image of the map
P\mapstoP(a)
\Z[\sqrt{2}]=\{P(\sqrt{2})\midP(X)\in\Z[X]\}
\Z[\sqrt{2}]=\{a+b\sqrt2\mida\in\Z,b\in\Z\}.
If is a field, the polynomial ring has many properties that are similar to those of the ring of integers
\Z.
Most of the properties of that are listed in this section do not remain true if is not a field, or if one considers polynomials in several indeterminates.
Like for integers, the Euclidean division of polynomials has a property of uniqueness. That is, given two polynomials and in, there is a unique pair of polynomials such that, and either or . This makes a Euclidean domain. However, most other Euclidean domains (except integers) do not have any property of uniqueness for the division nor an easy algorithm (such as long division) for computing the Euclidean division.
The Euclidean division is the basis of the Euclidean algorithm for polynomials that computes a polynomial greatest common divisor of two polynomials. Here, "greatest" means "having a maximal degree" or, equivalently, being maximal for the preorder defined by the degree. Given a greatest common divisor of two polynomials, the other greatest common divisors are obtained by multiplication by a nonzero constant (that is, all greatest common divisors of and are associated). In particular, two polynomials that are not both zero have a unique greatest common divisor that is monic (leading coefficient equal to).
The extended Euclidean algorithm allows computing (and proving) Bézout's identity. In the case of, it may be stated as follows. Given two polynomials and of respective degrees and, if their monic greatest common divisor has the degree, then there is a unique pair of polynomials such that
ap+bq=g,
\deg(a)\len-d, \deg(b)<m-d.
\deg(a)=n-d
Euclid's lemma applies to . That is, if divides, and is coprime with, then divides . Here, coprime means that the monic greatest common divisor is . Proof: By hypothesis and Bézout's identity, there are,, and such that and . So
c=c(ap+bq)=cap+aeq=a(cp+eq).
The unique factorization property results from Euclid's lemma. In the case of integers, this is the fundamental theorem of arithmetic. In the case of, it may be stated as: every non-constant polynomial can be expressed in a unique way as the product of a constant, and one or several irreducible monic polynomials; this decomposition is unique up to the order of the factors. In other terms is a unique factorization domain. If is the field of complex numbers, the fundamental theorem of algebra asserts that a univariate polynomial is irreducible if and only if its degree is one. In this case the unique factorization property can be restated as: every non-constant univariate polynomial over the complex numbers can be expressed in a unique way as the product of a constant, and one or several polynomials of the form ; this decomposition is unique up to the order of the factors. For each factor, is a root of the polynomial, and the number of occurrences of a factor is the multiplicity of the corresponding root.
See main article: Formal derivative and Derivation (differential algebra). The (formal) derivative of the polynomial
a0+a1X+a
n | |
nX |
a1+2a2X+ … +na
n-1 | |
nX |
.
The existence of the derivative is one of the main properties of a polynomial ring that is not shared with integers, and makes some computations easier on a polynomial ring than on integers.
See main article: Square-free polynomial.
See main article: Polynomial decomposition.
See main article: Polynomial factorization.
Except for factorization, all previous properties of are effective, since their proofs, as sketched above, are associated with algorithms for testing the property and computing the polynomials whose existence are asserted. Moreover these algorithms are efficient, as their computational complexity is a quadratic function of the input size.
The situation is completely different for factorization: the proof of the unique factorization does not give any hint for a method for factorizing. Already for the integers, there is no known algorithm running on a classical (non-quantum) computer for factorizing them in polynomial time. This is the basis of the RSA cryptosystem, widely used for secure Internet communications.
In the case of, the factors, and the methods for computing them, depend strongly on . Over the complex numbers, the irreducible factors (those that cannot be factorized further) are all of degree one, while, over the real numbers, there are irreducible polynomials of degree 2, and, over the rational numbers, there are irreducible polynomials of any degree. For example, the polynomial
X4-2
(X-\sqrt[4]2)(X+\sqrt[4]2)(X2+\sqrt2)
(X-\sqrt[4]2)(X+\sqrt[4]2)(X-i\sqrt[4]2)(X+i\sqrt[4]2)
The existence of a factorization algorithm depends also on the ground field. In the case of the real or complex numbers, Abel–Ruffini theorem shows that the roots of some polynomials, and thus the irreducible factors, cannot be computed exactly. Therefore, a factorization algorithm can compute only approximations of the factors. Various algorithms have been designed for computing such approximations, see Root finding of polynomials.
There is an example of a field such that there exist exact algorithms for the arithmetic operations of, but there cannot exist any algorithm for deciding whether a polynomial of the form
Xp-a
On the other hand, over the rational numbers and over finite fields, the situation is better than for integer factorization, as there are factorization algorithms that have a polynomial complexity. They are implemented in most general purpose computer algebra systems.
See main article: Minimal polynomial (field theory).
If is an element of an associative -algebra, the polynomial evaluation at is the unique algebra homomorphism from into that maps to and does not affect the elements of itself (it is the identity map on). It consists of substituting with in every polynomial. That is,
\varphi\left(amXm+amXm+ … +a1X+a0\right)=am\thetam+am\thetam+ … +a1\theta+a0.
The image of this evaluation homomorphism is the subalgebra generated by, which is necessarily commutative.If is injective, the subalgebra generated by is isomorphic to . In this case, this subalgebra is often denoted by . The notation ambiguity is generally harmless, because of the isomorphism.
If the evaluation homomorphism is not injective, this means that its kernel is a nonzero ideal, consisting of all polynomials that become zero when is substituted with . This ideal consists of all multiples of some monic polynomial, that is called the minimal polynomial of . The term minimal is motivated by the fact that its degree is minimal among the degrees of the elements of the ideal.
There are two main cases where minimal polynomials are considered.
In field theory and number theory, an element of an extension field of is algebraic over if it is a root of some polynomial with coefficients in . The minimal polynomial over of is thus the monic polynomial of minimal degree that has as a root. Because is a field, this minimal polynomial is necessarily irreducible over . For example, the minimal polynomial (over the reals as well as over the rationals) of the complex number is
X2+1
In linear algebra, the square matrices over form an associative -algebra of finite dimension (as a vector space). Therefore the evaluation homomorphism cannot be injective, and every matrix has a minimal polynomial (not necessarily irreducible). By Cayley–Hamilton theorem, the evaluation homomorphism maps to zero the characteristic polynomial of a matrix. It follows that the minimal polynomial divides the characteristic polynomial, and therefore that the degree of the minimal polynomial is at most .
In the case of, the quotient ring by an ideal can be built, as in the general case, as a set of equivalence classes. However, as each equivalence class contains exactly one polynomial of minimal degree, another construction is often more convenient.
Given a polynomial of degree, the quotient ring of by the ideal generated by can be identified with the vector space of the polynomials of degrees less than, with the "multiplication modulo " as a multiplication, the multiplication modulo consisting of the remainder under the division by of the (usual) product of polynomials. This quotient ring is variously denoted as
K[X]/pK[X],
K[X]/\langlep\rangle,
K[X]/(p),
K[X]/p.
The ring
K[X]/(p)
For example, the standard definition of the field of the complex numbers can be summarized by saying that it is the quotient ring
C=R[X]/(X2+1),
C
R.
Let be an algebraic element in a -algebra . By algebraic, one means that has a minimal polynomial . The first ring isomorphism theorem asserts that the substitution homomorphism induces an isomorphism of
K[X]/(p)
K[X]/(p).
The structure theorem for finitely generated modules over a principal ideal domain applies toK[''X''], when K is a field. This means that every finitely generated module over K[''X''] may be decomposed into a direct sum of a free module and finitely many modules of the form
K[X]/\left\langlePk\right\rangle
Given symbols
X1,...,Xn,
\alpha1 | |
X | |
1 |
…
\alphan | |
X | |
n |
0 … | |
X | |
1 |
0 | |
X | |
n |
=1.
The tuple of exponents is called the multidegree or exponent vector of the monomial. For a less cumbersome notation, the abbreviation
\alpha1 | |
X | |
1 |
…
\alphan | |
X | |
n |
\deg\alpha=
n | |
\sum | |
i=1 |
\alphai.
A polynomial in these indeterminates, with coefficients in a field, or more generally a ring, is a finite linear combination of monomials
p=\sum\alphap\alphaX\alpha
The set of polynomials in
X1,...,Xn,
K[X1,...,Xn],
K[X1,...,Xn]
K[X1,...,Xn]
Addition and scalar multiplication of polynomials are those of a vector space or free module equipped by a specific basis (here the basis of the monomials). Explicitly, let
p=\sum\alpha\inp\alphaX\alpha, q=\sum\beta\inq\betaX\beta,
The scalar multiplication of and a scalar
c\inK
cp=\sum\alpha\incp\alphaX\alpha.
The addition of and is
p+q=\sum\alpha\in(p\alpha+q\alpha)X\alpha,
p\alpha=0
\alpha\not\inI,
q\beta=0
\beta\not\inJ.
p\alpha+q\alpha=0
\alpha\inI\capJ,
The multiplication is
pq=\sum\gamma\in\left(\sum\alpha,p\alphaq\beta\right)X\gamma,
I+J
The verification of the axioms of an associative algebra is straightforward.
See main article: Algebraic expression. A polynomial expression is an expression built with scalars (elements of), indeterminates, and the operators of addition, multiplication, and exponentiation to nonnegative integer powers.
As all these operations are defined in
K[X1,...,Xn]
K[X1,...,Xn].
The distinction between a polynomial expression and the polynomial that it represents is relatively recent, and mainly motivated by the rise of computer algebra, where, for example, the test whether two polynomial expressions represent the same polynomial may be a nontrivial computation.
If is a commutative ring, the polynomial ring has the following universal property: for every commutative -algebra, and every -tuple of elements of, there is a unique algebra homomorphism from to that maps each
Xi
xi.
Xi
xi
As it is the case for every universal property, this characterizes the pair
(K[X1,...,Xn],(X1,...,Xn))
F:ALG\toSET
X\mapstoK[X]
POL:SET\toALG
The universal property of the polynomial ring means that and are adjoint functors. That is, there is a bijection
\operatorname{Hom}SET(X,\operatorname{F}(A))\cong\operatorname{Hom}ALG(K[X],A).
This may be expressed also by saying that polynomial rings are free commutative algebras, since they are free objects in the category of commutative algebras. Similarly, a polynomial ring with integer coefficients is the free commutative ring over its set of variables, since commutative rings and commutative algebras over the integers are the same thing.
A polynomial in
K[X1,\ldots,Xn]
Xn
K[X1,\ldots,Xn-1],
Xn,
\sum | |
(\alpha1,\ldots,\alphan)\inI |
c | |
\alpha1,\ldots,\alphan |
\alpha1 | |
X | |
1 |
…
\alphan | |
X | |
n |
=\sumi\left(\sum
(\alpha1,\ldots,\alphan-1)\mid(\alpha1,\ldots,\alphan-1,i)\inI |
c | |
\alpha1,\ldots,\alphan-1 |
\alpha1 | |
X | |
1 |
…
\alphan-1 | |
X | |
n-1 |
i, | |
\right)X | |
n |
This means that one has an algebra isomorphism
K[X1,\ldots,Xn]\cong(K[X1,\ldots,Xn-1])[Xn]
In other words, a multivariate polynomial ring can be considered as a univariate polynomial over a smaller polynomial ring. This is commonly used for proving properties of multivariate polynomial rings, by induction on the number of indeterminates.
The main such properties are listed below.
In this section, is a commutative ring, is a field, denotes a single indeterminate, and, as usual,
Z
K[X1,\ldots,Xn]
Z[X1,\ldots,Xn]
L[X],
K[X1,\ldots,Xn]
Z[X1,\ldots,Xn]
K[X1,\ldots,Xn]
Z[X1,\ldots,Xn]
\dimR[X]=1+\dimR,
\dim
\dimK[X1,\ldots,Xn]=n
\dimZ[X1,\ldots,Xn]=n+1.
\operatorname{gl}\dim
K[X1,\ldots,Xn]
Z[X1,\ldots,Xn]
\operatorname{gl}\dimZ[X1,\ldots,Xn]=n+1,
\operatorname{gl}\dimK[X1,\ldots,Xn]=n.
Polynomial rings in several variables over a field are fundamental in invariant theory and algebraic geometry. Some of their properties, such as those described above can be reduced to the case of a single indeterminate, but this is not always the case. In particular, because of the geometric applications, many interesting properties must be invariant under affine or projective transformations of the indeterminates. This often implies that one cannot select one of the indeterminates for a recurrence on the indeterminates.
Bézout's theorem, Hilbert's Nullstellensatz and Jacobian conjecture are among the most famous properties that are specific to multivariate polynomials over a field.
See main article: Hilbert's Nullstellensatz. The Nullstellensatz (German for "zero-locus theorem") is a theorem, first proved by David Hilbert, which extends to the multivariate case some aspects of the fundamental theorem of algebra. It is foundational for algebraic geometry, as establishing a strong link between the algebraic properties of
K[X1,\ldots,Xn]
The Nullstellensatz, has three main versions, each being a corollary of any other. Two of these versions are given below. For the third version, the reader is referred to the main article on the Nullstellensatz.
The first version generalizes the fact that a nonzero univariate polynomial has a complex zero if and only if it is not a constant. The statement is: a set of polynomials in
K[X1,\ldots,Xn]
The second version generalizes the fact that the irreducible univariate polynomials over the complex numbers are associate to a polynomial of the form
X-\alpha.
K[X1,\ldots,Xn]
\langleX1-\alpha1,\ldots,Xn-\alphan\rangle.
See main article: Bézout's theorem. Bézout's theorem may be viewed as a multivariate generalization of the version of the fundamental theorem of algebra that asserts that a univariate polynomial of degree has complex roots, if they are counted with their multiplicities.
In the case of bivariate polynomials, it states that two polynomials of degrees and in two variables, which have no common factors of positive degree, have exactly common zeros in an algebraically closed field containing the coefficients, if the zeros are counted with their multiplicity and include the zeros at infinity.
For stating the general case, and not considering "zero at infinity" as special zeros, it is convenient to work with homogeneous polynomials, and consider zeros in a projective space. In this context, a projective zero of a homogeneous polynomial
P(X0,\ldots,Xn)
(x0,\ldots,xn)
P(x0,\ldots,xn)=0
(x0,\ldots,xn)
(λx0,\ldots,λxn)
λ\inK.
Then, Bézout's theorem states: Given homogeneous polynomials of degrees
d1,\ldots,dn
d1 … dn.
See main article: Jacobian conjecture.
Polynomial rings can be generalized in a great many ways, including polynomial rings with generalized exponents, power series rings, noncommutative polynomial rings, skew polynomial rings, and polynomial rigs.
One slight generalization of polynomial rings is to allow for infinitely many indeterminates. Each monomial still involves only a finite number of indeterminates (so that its degree remains finite), and each polynomial is a still a (finite) linear combination of monomials. Thus, any individual polynomial involves only finitely many indeterminates, and any finite computation involving polynomials remains inside some subring of polynomials in finitely many indeterminates. This generalization has the same property of usual polynomial rings, of being the free commutative algebra, the only difference is that it is a free object over an infinite set.
One can also consider a strictly larger ring, by defining as a generalized polynomial an infinite (or finite) formal sum of monomials with a bounded degree. This ring is larger than the usual polynomial ring, as it includes infinite sums of variables. However, it is smaller than the ring of power series in infinitely many variables. Such a ring is used for constructing the ring of symmetric functions over an infinite set.
See main article: Monoid ring. A simple generalization only changes the set from which the exponents on the variable are drawn. The formulas for addition and multiplication make sense as long as one can add exponents: . A set for which addition makes sense (is closed and associative) is called a monoid. The set of functions from a monoid N to a ring R which are nonzero at only finitely many places can be given the structure of a ring known as R[''N''], the monoid ring of N with coefficients in R. The addition is defined component-wise, so that if, then for every n in N. The multiplication is defined as the Cauchy product, so that if, then for each n in N, cn is the sum of all aibj where i, j range over all pairs of elements of N which sum to n.
When N is commutative, it is convenient to denote the function a in R[''N''] as the formal sum:
\sumnanXn
and then the formulas for addition and multiplication are the familiar:
\left(\sumnanXn\right)+\left(\sumnbnXn\right)=\sumn\left(an+
n | |
b | |
n\right)X |
and
\left(\sumnanXn\right) ⋅ \left(\sumnbnXn\right)=\sumn\left(\sumi+j=nai
n | |
b | |
j\right)X |
where the latter sum is taken over all i, j in N that sum to n.
Some authors such as go so far as to take this monoid definition as the starting point, and regular single variable polynomials are the special case where N is the monoid of non-negative integers. Polynomials in several variables simply take N to be the direct product of several copies of the monoid of non-negative integers.
Several interesting examples of rings and groups are formed by taking N to be the additive monoid of non-negative rational numbers, . See also Puiseux series.
See main article: Formal power series.
Power series generalize the choice of exponent in a different direction by allowing infinitely many nonzero terms. This requires various hypotheses on the monoid N used for the exponents, to ensure that the sums in the Cauchy product are finite sums. Alternatively, a topology can be placed on the ring, and then one restricts to convergent infinite sums. For the standard choice of N, the non-negative integers, there is no trouble, and the ring of formal power series is defined as the set of functions from N to a ring R with addition component-wise, and multiplication given by the Cauchy product. The ring of power series can also be seen as the ring completion of the polynomial ring with respect to the ideal generated by .
See main article: Free algebra.
For polynomial rings of more than one variable, the products X⋅Y and Y⋅X are simply defined to be equal. A more general notion of polynomial ring is obtained when the distinction between these two formal products is maintained. Formally, the polynomial ring in n noncommuting variables with coefficients in the ring R is the monoid ring R[''N''], where the monoid N is the free monoid on n letters, also known as the set of all strings over an alphabet of n symbols, with multiplication given by concatenation. Neither the coefficients nor the variables need commute amongst themselves, but the coefficients and variables commute with each other.
Just as the polynomial ring in n variables with coefficients in the commutative ring R is the free commutative R-algebra of rank n, the noncommutative polynomial ring in n variables with coefficients in the commutative ring R is the free associative, unital R-algebra on n generators, which is noncommutative when n > 1.
See main article: Ore extension.
Other generalizations of polynomials are differential and skew-polynomial rings.
A differential polynomial ring is a ring of differential operators formed from a ring R and a derivation δ of R into R. This derivation operates on R, and will be denoted X, when viewed as an operator. The elements of R also operate on R by multiplication. The composition of operators is denoted as the usual multiplication. It follows that the relation may be rewrittenas
X ⋅ a=a ⋅ X+\delta(a).
This relation may be extended to define a skew multiplication between two polynomials in X with coefficients in R, which make them a noncommutative ring.
The standard example, called a Weyl algebra, takes R to be a (usual) polynomial ring k[''Y'' ], and δ to be the standard polynomial derivative
\tfrac{\partial}{\partialY}
The skew-polynomial ring is defined similarly for a ring R and a ring endomorphism f of R, by extending the multiplication from the relation X⋅r = f(r)⋅X to produce an associative multiplication that distributes over the standard addition. More generally, given a homomorphism F from the monoid N of the positive integers into the endomorphism ring of R, the formula X n⋅r = F(n)(r)⋅X n allows constructing a skew-polynomial ring. Skew polynomial rings are closely related to crossed product algebras.
The definition of a polynomial ring can be generalised by relaxing the requirement that the algebraic structure R be a field or a ring to the requirement that R only be a semifield or rig; the resulting polynomial structure/extension R[''X''] is a polynomial rig. For example, the set of all multivariate polynomials with natural number coefficients is a polynomial rig.