Skip to main content

Related Concepts

Polynomial
Polynomial
In mathematics, a polynomial is a mathematical expression consisting of indeterminates and coefficients, that involves only the operations of addition, subtraction, multiplication and exponentiation to nonnegative integer powers, and has a finite number of terms. An example of a polynomial of a single indeterminate x is x² − 4x + 7. An example with three indeterminates is x³ + 2xyz² − yz + 1. Polynomials appear in many areas of mathematics and science. For example, they are used to form polynomial equations, which encode a wide range of problems, from elementary word problems to complicated scientific problems; they are used to define polynomial functions, which appear in settings ranging from basic chemistry and physics to economics and social science; and they are used in calculus and numerical analysis to approximate other functions. In advanced mathematics, polynomials are used to construct polynomial rings and algebraic varieties, which are central concepts in algebra and algebraic geometry. The word polynomial joins two diverse roots: the Greek poly, meaning "many", and the Latin nomen, or "name". It was derived from the term binomial by replacing the Latin root bi- with the Greek poly-. That is, it means a sum of many terms. The word polynomial was first used in the 17th century. The x occurring in a polynomial is commonly called a variable or an indeterminate. When the polynomial is considered as an expression, x is a fixed symbol which does not have any value. However, when one considers the function defined by the polynomial, then x represents the argument of the function, and is therefore called a "variable". Many authors use these two words interchangeably. A polynomial P in the indeterminate x is commonly denoted either as P or as P. Formally, the name of the polynomial is P, not P, but the use of the functional notation P dates from a time when the distinction between a polynomial and the associated function was unclear. Moreover, the functional notation is often useful for specifying, in a single phrase, a polynomial and its indeterminate. For example, "let P be a polynomial" is a shorthand for "let P be a polynomial in the indeterminate x". On the other hand, when it is not necessary to emphasize the name of the indeterminate, many formulas are much simpler and easier to read if the name of the indeterminate do not appear at each occurrence of the polynomial. The ambiguity of having two notations for a single mathematical object may be formally resolved by considering the general meaning of the functional notation for polynomials. If a denotes a number, a variable, another polynomial, or, more generally, any expression, then P(a) denotes, by convention, the result of substituting a for x in P. Thus, the polynomial P defines the function displaystyleamapstoP(a), which is the polynomial function associated to P. Frequently, when using this notation, one supposes that a is a number. However, one may use it over any domain where addition and multiplication are defined (that is, any ring). In particular, if a is a polynomial then P(a) is also a polynomial. More specifically, when a is the indeterminate x, then the image of x by this function is the polynomial P itself (substituting x for x does not change anything). In other words, displaystyleP(x)=P, which justifies formally the existence of two notations for the same polynomial. A polynomial expression is an expression that can be built from constants and symbols called variables or indeterminates by means of addition, multiplication and exponentiation to a non-negative integer power. The constants are generally numbers, but may be any expression that do not involve the indeterminates, and represent mathematical objects that can be added and multiplied. Two polynomial expressions are considered as defining the same polynomial if they may be transformed, one to the other, by applying the usual properties of commutativity, associativity and distributivity of addition and multiplication. For example displaystyle(x-1)(x-2) and displaystylex²-3x+2 are two polynomial expressions that represent the same polynomial; so, one has the equality displaystyle(x-1)(x-2)=x²-3x+2. A polynomial in a single indeterminate x can always be written (or rewritten) in the form displaystyleaₙxⁿ+aₙ₋₁xⁿ⁻¹+dotsb+a₂x²+a₁x+a₀, where displaystylea₀,ldots,aₙ are constants that are called the coefficients of the polynomial, and displaystyle x is the indeterminate. The word "indeterminate" means that displaystyle x represents no particular value, although any value may be substituted for it. The mapping that associates the result of this substitution to the substituted value is a function, called a polynomial function. This can be expressed more concisely by using summation notation: displaystylesumₖ₌₀ⁿaₖxᵏ That is, a polynomial can either be zero or can be written as the sum of a finite number of non-zero terms. Each term consists of the product of a number – called the coefficient of the term – and a finite number of indeterminates, raised to non-negative integer powers. The exponent on an indeterminate in a term is called the degree of that indeterminate in that term; the degree of the term is the sum of the degrees of the indeterminates in that term, and the degree of a polynomial is the largest degree of any term with nonzero coefficient. Because x = x¹, the degree of an indeterminate without a written exponent is one. A term with no indeterminates and a polynomial with no indeterminates are called, respectively, a constant term and a constant polynomial. The degree of a constant term and of a nonzero constant polynomial is 0. The degree of the zero polynomial 0 is generally treated as not defined. For example: displaystyle-5x²y is a term. The coefficient is −5, the indeterminates are x and y, the degree of x is two, while the degree of y is one. The degree of the entire term is the sum of the degrees of each indeterminate in it, so in this example the degree is 2 + 1 = 3. Forming a sum of several terms produces a polynomial. For example, the following is a polynomial: displaystyleunderbrace,3x²bₑgᵢₙₛₘₐₗₗₘₐₜᵣᵢₓₘₐₜₕᵣₘₜₑᵣₘ₁ₑₙdₛₘₐₗₗₘₐₜᵣᵢₓunderbrace-,5xbₑgᵢₙₛₘₐₗₗₘₐₜᵣᵢₓₘₐₜₕᵣₘₜₑᵣₘ₂ₑₙdₛₘₐₗₗₘₐₜᵣᵢₓunderbrace+,4bₑgᵢₙₛₘₐₗₗₘₐₜᵣᵢₓₘₐₜₕᵣₘₜₑᵣₘ₃ₑₙdₛₘₐₗₗₘₐₜᵣᵢₓ. It consists of three terms: the first is degree two, the second is degree one, and the third is degree zero. Polynomials of small degree have been given specific names. A polynomial of degree zero is a constant polynomial, or simply a constant. Polynomials of degree one, two or three are respectively linear polynomials, quadratic polynomials and cubic polynomials. For higher degrees, the specific names are not commonly used, although quartic polynomial and quintic polynomial are sometimes used. The names for the degrees may be applied to the polynomial or to its terms. For example, the term 2x in x² + 2x + 1 is a linear term in a quadratic polynomial. The polynomial 0, which may be considered to have no terms at all, is called the zero polynomial. Unlike other constant polynomials, its degree is not zero. Rather, the degree of the zero polynomial is either left explicitly undefined, or defined as negative. The zero polynomial is also unique in that it is the only polynomial in one indeterminate that has an infinite number of roots. The graph of the zero polynomial, f = 0, is the x-axis. In the case of polynomials in more than one indeterminate, a polynomial is called homogeneous of degree n if all of its non-zero terms have degree n. The zero polynomial is homogeneous, and, as a homogeneous polynomial, its degree is undefined. For example, x³y² + 7x²y³ − 3x⁵ is homogeneous of degree 5. For more details, see Homogeneous polynomial. The commutative law of addition can be used to rearrange terms into any preferred order. In polynomials with one indeterminate, the terms are usually ordered according to degree, either in "descending powers of x", with the term of largest degree first, or in "ascending powers of x". The polynomial 3x² − 5x + 4 is written in descending powers of x. The first term has coefficient 3, indeterminate x, and exponent 2. In the second term, the coefficient is −5. The third term is a constant. Because the degree of a non-zero polynomial is the largest degree of any one term, this polynomial has degree two. Two terms with the same indeterminates raised to the same powers are called "similar terms" or "like terms", and they can be combined, using the distributive law, into a single term whose coefficient is the sum of the coefficients of the terms that were combined. It may happen that this makes the coefficient 0. Polynomials can be classified by the number of terms with nonzero coefficients, so that a one-term polynomial is called a monomial, a two-term polynomial is called a binomial, and a three-term polynomial is called a trinomial. A real polynomial is a polynomial with real coefficients. When it is used to define a function, the domain is not so restricted. However, a real polynomial function is a function from the reals to the reals that is defined by a real polynomial. Similarly, an integer polynomial is a polynomial with integer coefficients, and a complex polynomial is a polynomial with complex coefficients. A polynomial in one indeterminate is called a univariate polynomial, a polynomial in more than one indeterminate is called a multivariate polynomial. A polynomial with two indeterminates is called a bivariate polynomial. These notions refer more to the kind of polynomials one is generally working with than to individual polynomials; for instance, when working with univariate polynomials, one does not exclude constant polynomials, although strictly speaking, constant polynomials do not contain any indeterminates at all. It is possible to further classify multivariate polynomials as bivariate, trivariate, and so on, according to the maximum number of indeterminates allowed. Again, so that the set of objects under consideration be closed under subtraction, a study of trivariate polynomials usually allows bivariate polynomials, and so on. It is also common to say simply "polynomials in x, y, and z", listing the indeterminates allowed. Polynomials can be added using the associative law of addition (grouping all their terms together into a single sum), possibly followed by reordering (using the commutative law) and combining of like terms. For example, if displaystyleP=3x²-2x+5xy-2 and displaystyleQ=-3x²+3x+4y²+8 then the sum displaystyleP+Q=3x²-2x+5xy-2-3x²+3x+4y²+8 can be reordered and regrouped as displaystyleP+Q=(3x²-3x²)+(-2x+3x)+5xy+4y²+(8-2) and then simplified to displaystyleP+Q=x+5xy+4y²+6. When polynomials are added together, the result is another polynomial. Subtraction of polynomials is similar. Polynomials can also be multiplied. To expand the product of two polynomials into a sum of terms, the distributive law is repeatedly applied, which results in each term of one polynomial being multiplied by every term of the other. For example, if displaystylebeginalignedcolorRedP&colorRed=2x+3y+5BlueQ&colorBlue=2x+5y+xy+1endaligned then displaystylebeginarrayrccrcrcrcrcolorRedPcolorBlueQ&=&&(colorRed2xcdotcolorBlue2x)&+&(colorRed2xcdotcolorBlue5y)&+&(colorRed2xcdotcolorBluexy)&+&(colorRed2xcdotcolorBlue1)&&+&(colorRed3ycdotcolorBlue2x)&+&(colorRed3ycdotcolorBlue5y)&+&(colorRed3ycdotcolorBluexy)&+&(colorRed3ycdotcolorBlue1)&&+&(colorRed5cdotcolorBlue2x)&+&(colorRed5cdotcolorBlue5y)&+&(colorRed5cdotcolorBluexy)&+&(colorRed5cdotcolorBlue1)endarray Carrying out the multiplication in each term produces displaystylebeginarrayrccrcrcrcrPQ&=&&4x²&+&10xy&+&2x²y&+&2x&&+&6xy&+&15y²&+&3xy²&+&3y&&+&10x&+&25y&+&5xy&+&5.endarray Combining similar terms yields displaystylebeginarrayrcccrcrcrcrPQ&=&&4x²&+&(10xy+6xy+5xy)&+&2x²y&+&(2x+10x)&&+&15y²&+&3xy²&+&(3y+25y)&+&5endarray which can be simplified to displaystylePQ=4x²+21xy+2x²y+12x+15y²+3xy²+28y+5. As in the example, the product of polynomials is always a polynomial. Given a polynomial displaystyle f of a single variable and another polynomial g of any number of variables, the composition displaystyle fcirc g is obtained by substituting each copy of the variable of the first polynomial by the second polynomial. For example, if displaystylef(x)=x²+2x and displaystyleg(x)=3x+2 then displaystyle(fcirc g)(x)=f(g(x))=(3x+2)²+2(3x+2). A composition may be expanded to a sum of terms using the rules for multiplication and division of polynomials. The composition of two polynomials is another polynomial. The division of one polynomial by another is not typically a polynomial. Instead, such ratios are a more general family of objects, called rational fractions, rational expressions, or rational functions, depending on context. This is analogous to the fact that the ratio of two integers is a rational number, not necessarily an integer. For example, the fraction 1/ is not a polynomial, and it cannot be written as a finite sum of powers of the variable x. For polynomials in one variable, there is a notion of Euclidean division of polynomials, generalizing the Euclidean division of integers. This notion of the division a/b results in two polynomials, a quotient q and a remainder r, such that a = b q + r and degree < degree. The quotient and remainder may be computed by any of several algorithms, including polynomial long division and synthetic division. When the denominator b is monic and linear, that is, b = x − c for some constant c, then the polynomial remainder theorem asserts that the remainder of the division of a by b is the evaluation a. In this case, the quotient may be computed by Ruffini's rule, a special case of synthetic division. All polynomials with coefficients in a unique factorization domain (for example, the integers or a field) also have a factored form in which the polynomial is written as a product of irreducible polynomials and a constant. This factored form is unique up to the order of the factors and their multiplication by an invertible constant. In the case of the field of complex numbers, the irreducible factors are linear. Over the real numbers, they have the degree either one or two. Over the integers and the rational numbers the irreducible factors may have any degree. For example, the factored form of displaystyle5x³-5 is displaystyle5(x-1)left(x²+x+1right) over the integers and the reals, and displaystyle5(x-1)left(x+frac1+isqrt32right)left(x+frac1-isqrt32right) over the complex numbers. The computation of the factored form, called factorization is, in general, too difficult to be done by hand-written computation. However, efficient polynomial factorization algorithms are available in most computer algebra systems. Calculating derivatives and integrals of polynomials is particularly simple, compared to other kinds of functions. The derivative of the polynomial displaystyleP=aₙxⁿ+aₙ₋₁xⁿ⁻¹+dots+a₂x²+a₁x+a₀=sumᵢ₌₀ⁿaᵢxⁱ with respect to x is the polynomial displaystylenaₙxⁿ⁻¹+(n-1)aₙ₋₁xⁿ⁻²+dots+2a₂x+a₁=sumᵢ₌₁ⁿiaᵢxⁱ⁻¹. Similarly, the general antiderivative (or indefinite integral) of displaystyle P is displaystylefracaₙxⁿ⁺¹n+1+fracaₙ₋₁xⁿn+dots+fraca₂x³3+fraca₁x²2+a₀x+c=c+sumᵢ₌₀ⁿfracaᵢxⁱ⁺¹i+1 where c is an arbitrary constant. For example, antiderivatives of x² + 1 have the form ⁠1/3⁠x³ + x + c. For polynomials whose coefficients come from more abstract settings, the formula for the derivative can still be interpreted formally, with the coefficient kaₖ understood to mean the sum of k copies of aₖ. For example, over the integers modulo p, the derivative of the polynomial x + x is the polynomial 1. A polynomial function is a function that can be defined by evaluating a polynomial. More precisely, a function f of one argument from a given domain is a polynomial function if there exists a polynomial displaystyleaₙxⁿ+aₙ₋₁xⁿ⁻¹+cdots+a₂x²+a₁x+a₀ that evaluates to displaystylef(x) for all x in the domain of f (here, n is a non-negative integer and a₀, a₁, a₂, ..., aₙ are constant coefficients). Generally, unless otherwise specified, polynomial functions have complex coefficients, arguments, and values. In particular, a polynomial, restricted to have real coefficients, defines a function from the complex numbers to the complex numbers. If the domain of this function is also restricted to the reals, the resulting function is a real function that maps reals to reals. For example, the function f, defined by displaystylef(x)=x³-x, is a polynomial function of one variable. Polynomial functions of several variables are similarly defined, using polynomials in more than one indeterminate, as in displaystylef(x,y)=2x³+4x²y+xy⁵+y²-7. According to the definition of polynomial functions, there may be expressions that obviously are not polynomials but nevertheless define polynomial functions. An example is the expression displaystyleleft(sqrt1-x²right)², which takes the same values as the polynomial displaystyle1-x² on the interval displaystyle[-1,1], and thus both expressions define the same polynomial function on this interval. Every polynomial function is continuous, smooth, and entire. The evaluation of a polynomial is the computation of the corresponding polynomial function; that is, the evaluation consists of substituting a numerical value to each indeterminate and carrying out the indicated multiplications and additions. For polynomials in one indeterminate, the evaluation is usually more efficient (lower number of arithmetic operations to perform) using Horner's method, which consists of rewriting the polynomial as displaystyle(((((aₙx+aₙ₋₁)x+aₙ₋₂)x+dotsb+a₃)x+a₂)x+a₁)x+a₀. A polynomial function in one real variable can be represented by a graph. The graph of the zero polynomial f = 0 is the x-axis. The graph of a degree 0 polynomial f = a₀, where a₀ ≠ 0, is a horizontal line with y-intercept a₀ The graph of a degree 1 polynomial f = a₀ + a₁x, where a₁ ≠ 0, is an oblique line with y-intercept a₀ and slope a₁. The graph of a degree 2 polynomial f = a₀ + a₁x + a₂x², where a₂ ≠ 0 is a parabola. The graph of a degree 3 polynomial f = a₀ + a₁x + a₂x² + a₃x³, where a₃ ≠ 0 is a cubic curve. The graph of any polynomial with degree 2 or greater f = a₀ + a₁x + a₂x² + ⋯ + aₙx, where aₙ ≠ 0 and n ≥ 2 is a continuous non-linear curve. A non-constant polynomial function tends to infinity when the variable increases indefinitely. If the degree is higher than one, the graph does not have any asymptote. It has two parabolic branches with vertical direction. Polynomial graphs are analyzed in calculus using intercepts, slopes, concavity, and end behavior. A polynomial equation, also called an algebraic equation, is an equation of the form displaystyleaₙxⁿ+aₙ₋₁xⁿ⁻¹+dotsb+a₂x²+a₁x+a₀=0. For example, displaystyle3x²+4x-5=0 is a polynomial equation. When considering equations, the indeterminates of polynomials are also called unknowns, and the solutions are the possible values of the unknowns for which the equality is true. A polynomial equation stands in contrast to a polynomial identity like = x² − y², where both expressions represent the same polynomial in different forms, and as a consequence any evaluation of both members gives a valid equality. In elementary algebra, methods such as the quadratic formula are taught for solving all first degree and second degree polynomial equations in one variable. There are also formulas for the cubic and quartic equations. For higher degrees, the Abel–Ruffini theorem asserts that there can not exist a general formula in radicals. However, root-finding algorithms may be used to find numerical approximations of the roots of a polynomial expression of any degree. The number of solutions of a polynomial equation with real coefficients may not exceed the degree, and equals the degree when the complex solutions are counted with their multiplicity. This fact is called the fundamental theorem of algebra. A root of a nonzero univariate polynomial P is a value a of x such that P(a) = 0. In other words, a root of P is a solution of the polynomial equation P = 0 or a zero of the polynomial function defined by P. In the case of the zero polynomial, every number is a zero of the corresponding function, and the concept of root is rarely considered. A number a is a root of a polynomial P if and only if the linear polynomial x − a divides P, that is if there is another polynomial Q such that P = Q. It may happen that a power of x − a divides P; in this case, a is a multiple root of P, and otherwise a is a simple root of P. If P is a nonzero polynomial, there is a highest power m such that divides P, which is called the multiplicity of a as a root of P. The number of roots of a nonzero polynomial P, counted with their respective multiplicities, cannot exceed the degree of P, and equals this degree if all complex roots are considered. The coefficients of a polynomial and its roots are related by Vieta's formulas. Some polynomials, such as x² + 1, do not have any roots among the real numbers. If, however, the set of accepted solutions is expanded to the complex numbers, every non-constant polynomial has at least one root; this is the fundamental theorem of algebra. By successively dividing out factors x − a, one sees that any polynomial with complex coefficients can be written as a constant times a product of such polynomial factors of degree 1; as a consequence, the number of roots counted with their multiplicities is exactly equal to the degree of the polynomial. There may be several meanings of "solving an equation". One may want to express the solutions as explicit numbers; for example, the unique solution of 2x − 1 = 0 is 1/2. This is, in general, impossible for equations of degree greater than one, and, since the ancient times, mathematicians have searched to express the solutions as algebraic expressions; for example, the golden ratio displaystyle(1+sqrt5)/2 is the unique positive solution of displaystylex²-x-1=0. In the ancient times, they succeeded only for degrees one and two. For quadratic equations, the quadratic formula provides such expressions of the solutions. Since the 16th century, similar formulas (using cube roots in addition to square roots), although much more complicated, are known for equations of degree three and four (see cubic equation and quartic equation). But formulas for degree 5 and higher eluded researchers for several centuries. In 1824, Niels Henrik Abel proved the striking result that there are equations of degree 5 whose solutions cannot be expressed by a (finite) formula, involving only arithmetic operations and radicals (see Abel–Ruffini theorem). In 1830, Évariste Galois proved that most equations of degree higher than four cannot be solved by radicals, and showed that for each equation, one may decide whether it is solvable by radicals, and, if it is, solve it. This result marked the start of Galois theory and group theory, two important branches of modern algebra. Galois himself noted that the computations implied by his method were impracticable. Nevertheless, formulas for solvable equations of degrees 5 and 6 have been published (see quintic function and sextic equation). When there is no algebraic expression for the roots, and when such an algebraic expression exists but is too complicated to be useful, the unique way of solving it is to compute numerical approximations of the solutions. There are many methods for that; some are restricted to polynomials and others may apply to any continuous function. The most efficient algorithms allow solving easily polynomial equations of degree higher than 1,000. For polynomials with more than one indeterminate, the combinations of values for the variables for which the polynomial function takes the value zero are generally called zeros instead of "roots". The study of the sets of zeros of polynomials is the object of algebraic geometry. For a set of polynomial equations with several unknowns, there are algorithms to decide whether they have a finite number of complex solutions, and, if this number is finite, for computing the solutions. See System of polynomial equations. The special case where all the polynomials are of degree one is called a system of linear equations, for which another range of different solution methods exist, including the classical Gaussian elimination. A polynomial equation for which one is interested only in the solutions which are integers is called a Diophantine equation. Solving Diophantine equations is generally a very hard task. It has been proved that there cannot be any general algorithm for solving them, or even for deciding whether the set of solutions is empty. Some of the most famous problems that have been solved during the last fifty years are related to Diophantine equations, such as Fermat's Last Theorem. Polynomials where indeterminates are substituted for some other mathematical objects are often considered, and sometimes have a special name. A trigonometric polynomial is a finite linear combination of functions sin and cos with n taking on the values of one or more natural numbers. The coefficients may be taken as real numbers, for real-valued functions. If sin and cos are expanded in terms of sin and cos, a trigonometric polynomial becomes a polynomial in the two variables sin and cos. Conversely, every polynomial in sin and cos may be converted, with Product-to-sum identities, into a linear combination of functions sin and cos. This equivalence explains why linear combinations are called polynomials. For complex coefficients, there is no difference between such a function and a finite Fourier series. Trigonometric polynomials are widely used, for example in trigonometric interpolation applied to the interpolation of periodic functions. They are also used in the discrete Fourier transform. A matrix polynomial is a polynomial with square matrices as variables. Given an ordinary, scalar-valued polynomial displaystyleP(x)=sumᵢ₌₀ⁿaᵢxⁱ=a₀+a₁x+a₂x²+cdots+aₙxⁿ, this polynomial evaluated at a matrix A is displaystyleP(A)=sumᵢ₌₀ⁿaᵢAⁱ=a₀I+a₁A+a₂A²+cdots+aₙAⁿ, where I is the identity matrix. A matrix polynomial equation is an equality between two matrix polynomials, which holds for the specific matrices in question. A matrix polynomial identity is a matrix polynomial equation which holds for all matrices A in a specified matrix ring Mₙ. A bivariate polynomial where the second variable is substituted for an exponential function applied to the first variable, for example P, may be called an exponential polynomial. A rational fraction is the quotient of two polynomials. Any algebraic expression that can be rewritten as a rational fraction is a rational function. While polynomial functions are defined for all values of the variables, a rational function is defined only for the values of the variables for which the denominator is not zero. The rational fractions include the Laurent polynomials, but do not limit denominators to powers of an indeterminate. Laurent polynomials are like polynomials, but allow negative powers of the variable to occur. Formal power series are like polynomials, but allow infinitely many non-zero terms to occur, so that they do not have finite degree. Unlike polynomials they cannot in general be explicitly and fully written down, but the rules for manipulating their terms are the same as for polynomials. Non-formal power series also generalize polynomials, but the multiplication of two power series may not converge. A polynomial f over a commutative ring R is a polynomial all of whose coefficients belong to R. It is straightforward to verify that the polynomials in a given set of indeterminates over R form a commutative ring, called the polynomial ring in these indeterminates, denoted displaystyleR[x] in the univariate case and displaystyleR[x₁,ldots,xₙ] in the multivariate case. One has displaystyleR[x₁,ldots,xₙ]=left(R[x₁,ldots,xₙ₋₁]right)[xₙ]. So, most of the theory of the multivariate case can be reduced to an iterated univariate case. The map from R to R sending r to itself considered as a constant polynomial is an injective ring homomorphism, by which R is viewed as a subring of R. In particular, R is an algebra over R. One can think of the ring R as arising from R by adding one new element x to R, and extending in a minimal way to a ring in which x satisfies no other relations than the obligatory ones, plus commutation with all elements of R. To do this, one must add all powers of x and their linear combinations as well. Formation of the polynomial ring, together with forming factor rings by factoring out ideals, are important tools for constructing new rings out of known ones. For instance, the ring of complex numbers, which can be constructed from the polynomial ring R over the real numbers by factoring out the ideal of multiples of the polynomial x² + 1. Another example is the construction of finite fields, which proceeds similarly, starting out with the field of integers modulo some prime number as the coefficient ring R. If R is commutative, then one can associate with every polynomial P in R a polynomial function f with domain and range equal to R. One obtains the value f by substitution of the value r for the symbol x in P. One reason to distinguish between polynomials and polynomial functions is that, over some rings, different polynomials may give rise to the same polynomial function. This is not the case when R is the real or complex numbers, whence the two concepts are not always distinguished in analysis. An even more important reason to distinguish between polynomials and polynomial functions is that many operations on polynomials require looking at what a polynomial is composed of as an expression rather than evaluating it at some constant value for x. If R is an integral domain and f and g are polynomials in R[x], it is said that f divides g or f is a divisor of g if there exists a polynomial q in R[x] such that f q = g. If displaystyleainR, then a is a root of f if and only displaystylex-a divides f. In this case, the quotient can be computed using the polynomial long division. If F is a field and f and g are polynomials in F[x] with g ≠ 0, then there exist unique polynomials q and r in F[x] with displaystylef=q,g+r and such that the degree of r is smaller than the degree of g (using the convention that the polynomial 0 has a negative degree). The polynomials q and r are uniquely determined by f and g. This is called Euclidean division, division with remainder or polynomial long division and shows that the ring F[x] is a Euclidean domain. Analogously, prime polynomials can be defined as non-zero polynomials which cannot be factorized into the product of two non-constant polynomials. In the case of coefficients in a ring, "non-constant" must be replaced by "non-constant or non-unit". Any polynomial may be decomposed into the product of an invertible constant by a product of irreducible polynomials. If the coefficients belong to a field or a unique factorization domain this decomposition is unique up to the order of the factors and the multiplication of any non-unit factor by a unit. When the coefficients belong to integers, rational numbers or a finite field, there are algorithms to test irreducibility and to compute the factorization into irreducible polynomials. These algorithms are not practicable for hand-written computation, but are available in any computer algebra system. Eisenstein's criterion can also be used in some cases to determine irreducibility. In modern positional numbers systems, such as the decimal system, the digits and their positions in the representation of an integer, for example, 45, are a shorthand notation for a polynomial in the radix or base, in this case, 4 × 10¹ + 5 × 10⁰. As another example, in radix 5, a string of digits such as 132 denotes the number 1 × 5² + 3 × 5¹ + 2 × 5⁰ = 42. This representation is unique. Let b be a positive integer greater than 1. Then every positive integer a can be expressed uniquely in the form displaystylea=rₘbᵐ+rₘ₋₁bᵐ⁻¹+dotsb+r₁b+r₀, where m is a nonnegative integer and the r's are integers such that 0 < rₘ < b and 0 ≤ rᵢ < b for i = 0, 1,..., m − 1. The simple structure of polynomial functions makes them quite useful in analyzing general functions using polynomial approximations. An important example in calculus is Taylor's theorem, which roughly states that every differentiable function locally looks like a polynomial function, and the Stone–Weierstrass theorem, which states that every continuous function defined on a compact interval of the real axis can be approximated on the whole interval as closely as desired by a polynomial function. Practical methods of approximation include polynomial interpolation and the use of splines. Polynomials are frequently used to encode information about some other object. The characteristic polynomial of a matrix or linear operator contains information about the operator's eigenvalues. The minimal polynomial of an algebraic element records the simplest algebraic relation satisfied by that element. The chromatic polynomial of a graph counts the number of proper colourings of that graph. The term "polynomial", as an adjective, can also be used for quantities or functions that can be written in polynomial form. For example, in computational complexity theory the phrase polynomial time means that the time it takes to complete an algorithm is bounded by a polynomial function of some variable, such as the size of the input. Determining the roots of polynomials, or "solving algebraic equations", is among the oldest problems in mathematics. However, the elegant and practical notation we use today only developed beginning in the 15th century. Before that, equations were written out in words. For example, an algebra problem from the Chinese Arithmetic in Nine Sections, c. 200 BCE, begins "Three sheafs of good crop, two sheafs of mediocre crop, and one sheaf of bad crop are sold for 29 dou." We would write 3x + 2y + z = 29. The earliest known use of the equal sign is in Robert Recorde's The Whetstone of Witte, 1557. The signs + for addition, − for subtraction, and the use of a letter for an unknown appear in Michael Stifel's Arithemetica integra, 1544. René Descartes, in La géometrie, 1637, introduced the concept of the graph of a polynomial equation. He popularized the use of letters from the beginning of the alphabet to denote constants and letters from the end of the alphabet to denote variables, as can be seen above, in the general formula for a polynomial in one variable, where the as denote constants and x denotes a variable. Descartes introduced the use of superscripts to denote exponents as well. List of polynomial topics Markushevich, A.I. , "Polynomial", Encyclopedia of Mathematics, EMS Press "Euler's Investigations on the Roots of Equations". Archived from the original on September 24, 2012.
Quadratic formula
Quadratic formula
In elementary algebra, the quadratic formula is a closed-form expression describing the solutions of a quadratic equation. Other ways of solving quadratic equations, such as completing the square, yield the same solutions. Given a general quadratic equation of the form ⁠displaystyletextstyleax²+bx+c=0⁠, with ⁠displaystyle x⁠ representing an unknown, and coefficients ⁠displaystyle a⁠, ⁠displaystyle b⁠, and ⁠displaystyle c⁠ representing known real or complex numbers with ⁠displaystyleaneq0⁠, the values of ⁠displaystyle x⁠ satisfying the equation, called the roots or zeros, can be found using the quadratic formula, displaystylex=frac-bpmsqrtb²-4ac2a, where the plus–minus symbol "⁠displaystylepm⁠" indicates that the equation has two roots. Written separately, these are: displaystylex₁=frac-b+sqrtb²-4ac2a,qquadx₂=frac-b-sqrtb²-4ac2a. The quantity ⁠displaystyletextstyleDelta=b²-4ac⁠ is known as the discriminant of the quadratic equation. If the coefficients ⁠displaystyle a⁠, ⁠displaystyle b⁠, and ⁠displaystyle c⁠ are real numbers then when ⁠displaystyleDelta>0⁠, the equation has two distinct real roots; when ⁠displaystyleDelta=0⁠, the equation has one repeated real root; and when ⁠displaystyleDelta<0⁠, the equation has no real roots but has two distinct complex roots, which are complex conjugates of each other. Geometrically, the roots represent the ⁠displaystyle x⁠ values at which the graph of the quadratic function ⁠displaystyletextstyley=ax²+bx+c⁠, a parabola, crosses the ⁠displaystyle x⁠-axis: the graph's ⁠displaystyle x⁠-intercepts. The quadratic formula can also be used to identify the parabola's axis of symmetry. The standard way to derive the quadratic formula is to apply the method of completing the square to the generic quadratic equation ⁠displaystyletextstyleax²+bx+c=0⁠. The idea is to manipulate the equation into the form ⁠displaystyletextstyle(x+k)²=s⁠ for some expressions ⁠displaystyle k⁠ and ⁠displaystyle s⁠ written in terms of the coefficients; take the square root of both sides; and then isolate ⁠displaystyle x⁠. We start by dividing the equation by the quadratic coefficient ⁠displaystyle a⁠, which is allowed because ⁠displaystyle a⁠ is non-zero. Afterwards, we subtract the constant term ⁠displaystylec/a⁠ to isolate it on the right-hand side: displaystylebeginalignedax²ᵛᵖʰᵃⁿᵗᵒᵐ|+bx+c&=0[3mu]x²+fracbax+fracca&=0[3mu]x²+fracbax&=-fracca.endaligned The left-hand side is now of the form ⁠displaystyletextstylex²+2kx⁠, and we can "complete the square" by adding a constant ⁠displaystyletextstylek²⁠ to obtain a squared binomial ⁠displaystyletextstylex²+2kx+k²=⁠⁠displaystyletextstyle(x+k)²⁠. In this example we add ⁠displaystyletextstyle(b/2a)²⁠ to both sides so that the left-hand side can be factored (see the figure): displaystylebeginalignedx²+2left(fracb2aright)x+left(fracb2aright)²&=-fracca+left(fracb2aright)²[5mu]left(x+fracb2aright)²&=fracb²-4ac4a².endaligned Because the left-hand side is now a perfect square, we can easily take the square root of both sides: displaystylex+fracb2a=pmfracsqrtb²-4ac2a. Finally, subtracting ⁠displaystyleb/2a⁠ from both sides to isolate ⁠displaystyle x⁠ produces the quadratic formula: displaystylex=frac-bpmsqrtb²-4ac2a. The quadratic formula can equivalently be written using various alternative expressions, for instance displaystylex=-fracb2apmsqrtleft(fracb2aright)²-fracca, which can be derived by first dividing a quadratic equation by ⁠displaystyle2a⁠, resulting in ⁠displaystyletextstyletfrac12x²+tfracb2ax+tfracc2a=0⁠, then substituting the new coefficients into the standard quadratic formula. Because this variant allows re-use of the intermediately calculated quantity ⁠displaystyletfracb2a⁠, it can slightly reduce the arithmetic involved. A lesser known quadratic formula, first mentioned by Giulio Fagnano, describes the same roots via an equation with the square root in the denominator (assuming ⁠displaystylecneq0⁠): displaystylex=frac2c-bmpsqrtb²-4ac. Here the minus–plus symbol "⁠displaystylemp⁠" indicates that the two roots of the quadratic equation, in the same order as the standard quadratic formula, are displaystylex₁=frac2c-b-sqrtb²-4ac,qquadx₂=frac2c-b+sqrtb²-4ac. This variant has been jokingly called the "citardauq" formula. When ⁠displaystyle-b⁠ has the opposite sign as either ⁠displaystyletextstyle+sqrtb²-4ac⁠ or ⁠displaystyletextstyle-sqrtb²-4ac⁠, subtraction can cause catastrophic cancellation, resulting in poor accuracy in numerical calculations; choosing between the version of the quadratic formula with the square root in the numerator or denominator depending on the sign of ⁠displaystyle b⁠ can avoid this problem. See § Numerical calculation below. This version of the quadratic formula is used in Muller's method for finding the roots of general functions. It can be derived from the standard formula from the identity ⁠displaystylex₁x₂=c/a⁠, one of Vieta's formulas. Alternately, it can be derived by dividing each side of the equation ⁠displaystyletextstyleax²+bx+c=0⁠ by ⁠displaystyletextstylex²⁠ to get ⁠displaystyletextstylecx⁻²+bx⁻¹+a=0⁠, applying the standard formula to find the two roots ⁠displaystyletextstylex⁻¹!⁠, and then taking the reciprocal to find the roots ⁠displaystyle x⁠ of the original equation. Any generic method or algorithm for solving quadratic equations can be applied to an equation with symbolic coefficients and used to derive some closed-form expression equivalent to the quadratic formula. Alternative methods are sometimes simpler than completing the square, and may offer interesting insight into other areas of mathematics. Instead of dividing by ⁠displaystyle a⁠ to isolate ⁠displaystyletextstylex²!⁠, it can be slightly simpler to multiply by ⁠displaystyle4a⁠ instead to produce ⁠displaystyletextstyle(2ax)²!⁠, which allows us to complete the square without need for fractions. Then the steps of the derivation are: Multiply each side by ⁠displaystyle4a⁠. Add ⁠displaystyletextstyleb²-4ac⁠ to both sides to complete the square. Take the square root of both sides. Isolate ⁠displaystyle x⁠. Applying this method to a generic quadratic equation with symbolic coefficients yields the quadratic formula: displaystylebeginalignedax²+bx+c&=0[3mu]4a²x²+4abx+4ac&=0[3mu]4a²x²+4abx+b²&=b²-4ac[3mu](2ax+b)²&=b²-4ac[3mu]2ax+b&=pmsqrtb²-4ac[5mu]x&=dfrac-bpmsqrtb²-4ac2a.vphantombigg)endaligned This method for completing the square is ancient and was known to the 8th–9th century Indian mathematician Śrīdhara. Compared with the modern standard method for completing the square, this alternate method avoids fractions until the last step and hence does not require a rearrangement after step 3 to obtain a common denominator in the right side. Another derivation uses a change of variables to eliminate the linear term. Then the equation takes the form ⁠displaystyletextstyleu²=s⁠ in terms of a new variable ⁠displaystyle u⁠ and some constant expression ⁠displaystyle s⁠, whose roots are then ⁠displaystyleu=pmsqrts⁠. By substituting ⁠displaystylex=u-tfracb2a⁠ into ⁠displaystyletextstyleax²+bx+c=0⁠, expanding the products and combining like terms, and then solving for ⁠displaystyletextstyleu²!⁠, we have: displaystylebeginalignedaleft(u-fracb2aright)²+bleft(u-fracb2aright)+c&=0[5mu]aleft(u²-fracbau+fracb²4a²right)+bleft(u-fracb2aright)+c&=0[5mu]au²-bu+fracb²4a+bu-fracb²2a+c&=0[5mu]au²+frac4ac-b²4a&=0[5mu]u²&=fracb²-4ac4a².endaligned Finally, after taking a square root of both sides and substituting the resulting expression for ⁠displaystyle u⁠ back into ⁠displaystylex=u-tfracb2a,⁠ the familiar quadratic formula emerges: displaystylex=frac-bpmsqrtb²-4ac2a. The following method was used by many historical mathematicians: Let the roots of the quadratic equation ⁠displaystyletextstyleax²+bx+c=0⁠ be ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠. The derivation starts from an identity for the square of a difference (valid for any two complex numbers), of which we can take the square root on both sides: displaystylebeginaligned(alpha-beta)²&=(alpha+beta)²-4alphabeta[3mu]alpha-beta&=pmsqrt(alpha+beta)²-4alphabeta.endaligned Since the coefficient ⁠displaystyleaneq0⁠, we can divide the quadratic equation by ⁠displaystyle a⁠ to obtain a monic polynomial with the same roots. Namely, displaystylex²+fracbax+fracca=(x-alpha)(x-beta)=x²-(alpha+beta)x+alphabeta. This implies that the sum ⁠displaystylealpha+beta=-tfracba⁠ and the product ⁠displaystylealphabeta=tfracca⁠. Thus the identity can be rewritten: displaystylealpha-beta=pmsqrtleft(-fracbaright)²-4fracca=pmfracsqrtb²-4aca. Therefore, displaystylebeginalignedalpha&=tfrac12(alpha+beta)+tfrac12(alpha-beta)=-fracb2apmfracsqrtb²-4ac2a,[10mu]beta&=tfrac12(alpha+beta)-tfrac12(alpha-beta)=-fracb2ampfracsqrtb²-4ac2a.endaligned The two possibilities for each of ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠ are the same two roots in opposite order, so we can combine them into the standard quadratic equation: displaystylex=frac-bpmsqrtb²-4ac2a. An alternative way of deriving the quadratic formula is via the method of Lagrange resolvents, which is an early part of Galois theory. This method can be generalized to give the roots of cubic polynomials and quartic polynomials, and leads to Galois theory, which allows one to understand the solution of algebraic equations of any degree in terms of the symmetry group of their roots, the Galois group. This approach focuses on the roots themselves rather than algebraically rearranging the original equation. Given a monic quadratic polynomial ⁠displaystyletextstylex²+px+q⁠ assume that ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠ are the two roots. So the polynomial factors as displaystylebeginalignedx²+px+q&=(x-alpha)(x-beta)[3mu]&=x²-(alpha+beta)x+alphabetaendaligned which implies ⁠displaystylep=-(alpha+beta)⁠ and ⁠displaystyleq=alphabeta⁠. Since multiplication and addition are both commutative, exchanging the roots ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠ will not change the coefficients ⁠displaystyle p⁠ and ⁠displaystyle q⁠: one can say that ⁠displaystyle p⁠ and ⁠displaystyle q⁠ are symmetric polynomials in ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠. Specifically, they are the elementary symmetric polynomials – any symmetric polynomial in ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠ can be expressed in terms of ⁠displaystylealpha+beta⁠ and ⁠displaystylealphabeta⁠ instead. The Galois theory approach to analyzing and solving polynomials is to ask whether, given coefficients of a polynomial each of which is a symmetric function in the roots, one can "break" the symmetry and thereby recover the roots. Using this approach, solving a polynomial of degree ⁠displaystyle n⁠ is related to the ways of rearranging ("permuting") ⁠displaystyle n⁠ terms, called the symmetric group on ⁠displaystyle n⁠ letters and denoted ⁠displaystyleSₙ⁠. For the quadratic polynomial, the only ways to rearrange two roots are to either leave them be or to transpose them, so solving a quadratic polynomial is simple. To find the roots ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠, consider their sum and difference: displaystyler₁=alpha+beta,quadr₂=alpha-beta. These are called the Lagrange resolvents of the polynomial, from which the roots can be recovered as displaystylealpha=tfrac12(r₁+r₂),quadbeta=tfrac12(r₁-r₂). Because ⁠displaystyler₁=alpha+beta⁠ is a symmetric function in ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠, it can be expressed in terms of ⁠displaystyle p⁠ and ⁠displaystyleq,⁠ specifically ⁠displaystyler₁=-p⁠ as described above. However, ⁠displaystyler₂=alpha-beta⁠ is not symmetric, since exchanging ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠ yields the additive inverse ⁠displaystyle-r₂=beta-alpha⁠. So ⁠displaystyler₂⁠ cannot be expressed in terms of the symmetric polynomials. However, its square ⁠displaystyletextstyler₂²=(alpha-beta)²⁠ is symmetric in the roots, expressible in terms of ⁠displaystyle p⁠ and ⁠displaystyle q⁠. Specifically ⁠displaystyletextstyler₂²=(alpha-beta)²=⁠⁠displaystyletextstyle(alpha+beta)²-4alphabeta=⁠⁠displaystyletextstylep²-4q⁠, which implies ⁠displaystyletextstyler₂=pmsqrtp²-4q⁠. Taking the positive root "breaks" the symmetry, resulting in displaystyler₁=-p,qquadr₂=textstylesqrtp²-4q from which the roots ⁠displaystylealpha⁠ and ⁠displaystylebeta⁠ are recovered as displaystylex=tfrac12(r₁pmr₂)=tfrac12bigl(-ppmtextstylesqrtp²-4q,bigr) which is the quadratic formula for a monic polynomial. Substituting ⁠displaystylep=b/a⁠, ⁠displaystyleq=c/a⁠ yields the usual expression for an arbitrary quadratic polynomial. The resolvents can be recognized as displaystyletfrac12r₁=-tfrac12p=-fracb2a,qquadr₂²=p₂-4q=fracb²-4aca², respectively the vertex and the discriminant of the monic polynomial. A similar but more complicated method works for cubic equations, which have three resolvents and a quadratic equation (the "resolving polynomial") relating ⁠displaystyler₂⁠ and ⁠displaystyler₃⁠, which one can solve by the quadratic equation, and similarly for a quartic equation (degree 4), whose resolving polynomial is a cubic, which can in turn be solved. The same method for a quintic equation yields a polynomial of degree 24, which does not simplify the problem, and, in fact, solutions to quintic equations in general cannot be expressed using only roots. The quadratic formula is exactly correct when performed using the idealized arithmetic of real numbers, but when approximate arithmetic is used instead, for example pen-and-paper arithmetic carried out to a fixed number of decimal places or the floating-point binary arithmetic available on computers, the limitations of the number representation can lead to substantially inaccurate results unless great care is taken in the implementation. Specific difficulties include catastrophic cancellation in computing the sum ⁠displaystyletextstyle-bpmsqrtDelta⁠ if ⁠displaystyletextstylebapproxpmsqrtDelta⁠; catastrophic calculation in computing the discriminant ⁠displaystyletextstyleDelta=b²-4ac⁠ itself in cases where ⁠displaystyletextstyleb²approx4ac⁠; degeneration of the formula when ⁠displaystyle a⁠, ⁠displaystyle b⁠, or ⁠displaystyle c⁠ is represented as zero or infinite; and possible overflow or underflow when multiplying or dividing extremely large or small numbers, even in cases where the roots can be accurately represented. Catastrophic cancellation occurs when two numbers which are approximately equal are subtracted. While each of the numbers may independently be representable to a certain number of digits of precision, the identical leading digits of each number cancel, resulting in a difference of lower relative precision. When ⁠displaystyletextstylebapproxsqrtDelta⁠, evaluation of ⁠displaystyletextstyle-b+sqrtDelta⁠ causes catastrophic cancellation, as does the evaluation of ⁠displaystyletextstyle-b-sqrtDelta⁠ when ⁠displaystyletextstylebapprox-sqrtDelta⁠. When using the standard quadratic formula, calculating one of the two roots always involves addition, which preserves the working precision of the intermediate calculations, while calculating the other root involves subtraction, which compromises it. Therefore, naïvely following the standard quadratic formula often yields one result with less relative precision than expected. Unfortunately, introductory algebra textbooks typically do not address this problem, even though it causes students to obtain inaccurate results in other school subjects such as introductory chemistry. For example, if trying to solve the equation ⁠displaystyletextstylex²-1634x+2=0⁠ using a pocket calculator, the result of the quadratic formula ⁠displaystyletextstylex=817pmsqrt667,487⁠ might be approximately calculated as: displaystylebeginalignedat3x₁&=817+816.998,776,0&&=1.633,998,776times10³,\x₂&=817-816.998,776,0&&=1.224times10⁻³.endalignedat Even though the calculator used ten decimal digits of precision for each step, calculating the difference between two approximately equal numbers has yielded a result for ⁠displaystylex₂⁠ with only four correct digits. One way to recover an accurate result is to use the identity ⁠displaystylex₁x₂=c/a⁠. In this example ⁠displaystylex₂⁠ can be calculated as ⁠displaystylex₂=2/x₁=⁠⁠displaystyle1.223,991,125times10⁻³!⁠, which is correct to the full ten digits. Another more or less equivalent approach is to use the version of the quadratic formula with the square root in the denominator to calculate one of the roots (see § Square root in the denominator above). Practical computer implementations of the solution of quadratic equations commonly choose which formula to use for each root depending on the sign of ⁠displaystyle b⁠. These methods do not prevent possible overflow or underflow of the floating-point exponent in computing ⁠displaystyletextstyleb²⁠ or ⁠displaystyle4ac⁠, which can lead to numerically representable roots not being computed accurately. A more robust but computationally expensive strategy is to start with the substitution ⁠displaystyletextstylex=-uoperatornamesgn(b)sqrtvertcvertbig/!sqrtvertavert⁠, turning the quadratic equation into displaystyleu²-2frac|b|2sqrt|a|sqrt|c|u+operatornamesgn(c)=0, where ⁠displaystyleoperatornamesgn⁠ is the sign function. Letting ⁠displaystyletextstyled=vertbvertbig/2sqrtvertavertsqrtvertcvert⁠, this equation has the form ⁠displaystyletextstyleu²-2dupm1=0⁠, for which one solution is ⁠displaystyletextstyleu₁=d+sqrtd²mp1⁠ and the other solution is ⁠displaystyletextstyleu₂=pm1/u₁⁠. The roots of the original equation are then ⁠displaystyletextstylex₁=-operatornamesgn(b)bigl(sqrtvertcvertbig/!sqrtvertavert !bigr)u₁⁠ and ⁠displaystyletextstylex₂=-operatornamesgn(b)bigl(sqrtvertcvertbig/!sqrtvertavert !bigr)u₂⁠. With additional complication the expense and extra rounding of the square roots can be avoided by approximating them as powers of two, while still avoiding exponent overflow for representable roots. The earliest methods for solving quadratic equations were geometric. Babylonian cuneiform tablets contain problems reducible to solving quadratic equations. The Egyptian Berlin Papyrus, dating back to the Middle Kingdom, contains the solution to a two-term quadratic equation. The Greek mathematician Euclid used geometric methods to solve quadratic equations in Book 2 of his Elements, an influential mathematical treatise Rules for quadratic equations appear in the Chinese The Nine Chapters on the Mathematical Art circa 200 BC. In his work Arithmetica, the Greek mathematician Diophantus solved quadratic equations with a method more recognizably algebraic than the geometric algebra of Euclid. His solution gives only one root, even when both roots are positive. The Indian mathematician Brahmagupta included a generic method for finding one root of a quadratic equation in his treatise Brāhmasphuṭasiddhānta (circa 628 AD), written out in words in the style of the time but more or less equivalent to the modern symbolic formula. His solution of the quadratic equation ⁠displaystyletextstyleax²+bx=c⁠ was as follows: "To the absolute number multiplied by four times the [coefficient of the] square, add the square of the [coefficient of the] middle term; the square root of the same, less the [coefficient of the] middle term, being divided by twice the [coefficient of the] square is the value." In modern notation, this can be written ⁠displaystyletextstylex=bigl(sqrtccdot4a+b²-bbigr)big/2a⁠. The Indian mathematician Śrīdhara (8th–9th century) came up with a similar algorithm for solving quadratic equations in a now-lost work on algebra quoted by Bhāskara II. The modern quadratic formula is sometimes called Sridharacharya's formula in India and Bhaskara's formula in Brazil. The 9th-century Persian mathematician Muḥammad ibn Mūsā al-Khwārizmī solved quadratic equations algebraically. The quadratic formula covering all cases was first obtained by Simon Stevin in 1594. In 1637 René Descartes published La Géométrie containing special cases of the quadratic formula in the form we know today. In terms of coordinate geometry, an axis-aligned parabola is a curve whose ⁠displaystyle(x,y)⁠-coordinates are the graph of a second-degree polynomial, of the form ⁠displaystyletextstyley=ax²+bx+c⁠, where ⁠displaystyle a⁠, ⁠displaystyle b⁠, and ⁠displaystyle c⁠ are real-valued constant coefficients with ⁠displaystyleaneq0⁠. Geometrically, the quadratic formula defines the points ⁠displaystyle(x,0)⁠ on the graph, where the parabola crosses the ⁠displaystyle x⁠-axis. Furthermore, it can be separated into two terms, displaystylex=frac-bpmsqrtb²-4ac2a=-fracb2apmfracsqrtb²-4ac2a. The first term describes the axis of symmetry, the line ⁠displaystylex=-tfracb2a⁠. The second term, ⁠displaystyletextstylesqrtb²-4acbig/2a⁠, gives the distance the roots are away from the axis of symmetry. If the parabola's vertex is on the ⁠displaystyle x⁠-axis, then the corresponding equation has a single repeated root on the line of symmetry, and this distance term is zero; algebraically, the discriminant ⁠displaystyletextstyleb²-4ac=0⁠. If the discriminant is positive, then the vertex is not on the ⁠displaystyle x⁠-axis but the parabola opens in the direction of the ⁠displaystyle x⁠-axis, crossing it twice, so the corresponding equation has two real roots. If the discriminant is negative, then the parabola opens in the opposite direction, never crossing the ⁠displaystyle x⁠-axis, and the equation has no real roots; in this case the two complex-valued roots will be complex conjugates whose real part is the ⁠displaystyle x⁠ value of the axis of symmetry. If the constants ⁠displaystyle a⁠, ⁠displaystyle b⁠, and/or ⁠displaystyle c⁠ are not unitless then the quantities ⁠displaystyle x⁠ and ⁠displaystyletfracba⁠ must have the same units, because the terms ⁠displaystyletextstyleax²⁠ and ⁠displaystyle bx⁠ agree on their units. By the same logic, the coefficient ⁠displaystyle c⁠ must have the same units as ⁠displaystyletfracb²a⁠, irrespective of the units of ⁠displaystyle x⁠. This can be a powerful tool for verifying that a quadratic expression of physical quantities has been set up correctly. Fundamental theorem of algebra Vieta's formulas Smith, David Eugene, History of Mathematics, vol.