Search results
Results from the WOW.Com Content Network
Like any series, an alternating series is a convergent series if and only if the sequence of partial sums of the series converges to a limit. The alternating series test guarantees that an alternating series is convergent if the terms a n converge to 0 monotonically, but this condition is not necessary for convergence.
The more general class of p-series, =, exemplifies the possible results of the test: If p ≤ 0, then the nth-term test identifies the series as divergent. If 0 < p ≤ 1, then the nth-term test is inconclusive, but the series is divergent by the integral test for convergence.
In mathematical analysis, the alternating series test proves that an alternating series is convergent when its terms decrease monotonically in absolute value and approach zero in the limit. The test was devised by Gottfried Leibniz and is sometimes known as Leibniz's test , Leibniz's rule , or the Leibniz criterion .
The rings for which such a theorem exists are called Euclidean domains, but in this generality, uniqueness of the quotient and remainder is not guaranteed. [8] Polynomial division leads to a result known as the polynomial remainder theorem: If a polynomial f(x) is divided by x − k, the remainder is the constant r = f(k). [9] [10]
This turns out to be equivalent to a system of simultaneous polynomial congruences, and may be solved by means of the Chinese remainder theorem for polynomials. Birkhoff interpolation is a further generalization where only derivatives of some orders are prescribed, not necessarily all orders from 0 to a k.
Thus, the function may be more "cheaply" evaluated using synthetic division and the polynomial remainder theorem. The factor theorem is another application of the remainder theorem: if the remainder is zero, then the linear divisor is a factor. Repeated application of the factor theorem may be used to factorize the polynomial. [3]
Conversely, the ratio of two alternating polynomials is a symmetric function, possibly rational (not necessarily a polynomial), though the ratio of an alternating polynomial over the Vandermonde polynomial is a polynomial. Schur polynomials are defined in this way, as an alternating polynomial divided by the Vandermonde polynomial.
The rule states that if the nonzero terms of a single-variable polynomial with real coefficients are ordered by descending variable exponent, then the number of positive roots of the polynomial is either equal to the number of sign changes between consecutive (nonzero) coefficients, or is less than it by an even number.