Riemann hypothesisIn mathematics, the Riemann hypothesis is the conjecture that the Riemann zeta function has its zeros only at the negative even integers and complex numbers with real part 1/2. Many consider it to be the most important unsolved problem in pure mathematics. It is of great interest in number theory because it implies results about the distribution of prime numbers. It was proposed by , after whom it is named.
Basel problemThe Basel problem is a problem in mathematical analysis with relevance to number theory, concerning an infinite sum of inverse squares. It was first posed by Pietro Mengoli in 1650 and solved by Leonhard Euler in 1734, and read on 5 December 1735 in The Saint Petersburg Academy of Sciences. Since the problem had withstood the attacks of the leading mathematicians of the day, Euler's solution brought him immediate fame when he was twenty-eight.
Convergent seriesIn mathematics, a series is the sum of the terms of an infinite sequence of numbers. More precisely, an infinite sequence defines a series S that is denoted The nth partial sum Sn is the sum of the first n terms of the sequence; that is, A series is convergent (or converges) if the sequence of its partial sums tends to a limit; that means that, when adding one after the other in the order given by the indices, one gets partial sums that become closer and closer to a given number.
Absolute convergenceIn mathematics, an infinite series of numbers is said to converge absolutely (or to be absolutely convergent) if the sum of the absolute values of the summands is finite. More precisely, a real or complex series is said to converge absolutely if for some real number Similarly, an improper integral of a function, is said to converge absolutely if the integral of the absolute value of the integrand is finite—that is, if Absolute convergence is important for the study of infinite series because its definition is strong enough to have properties of finite sums that not all convergent series possess – a convergent series that is not absolutely convergent is called conditionally convergent, while absolutely convergent series behave "nicely".
Harmonic numberIn mathematics, the n-th harmonic number is the sum of the reciprocals of the first n natural numbers: Starting from n = 1, the sequence of harmonic numbers begins: Harmonic numbers are related to the harmonic mean in that the n-th harmonic number is also n times the reciprocal of the harmonic mean of the first n positive integers. Harmonic numbers have been studied since antiquity and are important in various branches of number theory.
Analytic continuationIn complex analysis, a branch of mathematics, analytic continuation is a technique to extend the domain of definition of a given analytic function. Analytic continuation often succeeds in defining further values of a function, for example in a new region where the infinite series representation which initially defined the function becomes divergent. The step-wise continuation technique may, however, come up against difficulties. These may have an essentially topological nature, leading to inconsistencies (defining more than one value).
Series (mathematics)In mathematics, a series is, roughly speaking, the operation of adding infinitely many quantities, one after the other, to a given starting quantity. The study of series is a major part of calculus and its generalization, mathematical analysis. Series are used in most areas of mathematics, even for studying finite structures (such as in combinatorics) through generating functions. In addition to their ubiquity in mathematics, infinite series are also widely used in other quantitative disciplines such as physics, computer science, statistics and finance.
Riemann zeta functionThe Riemann zeta function or Euler–Riemann zeta function, denoted by the Greek letter ζ (zeta), is a mathematical function of a complex variable defined as for , and its analytic continuation elsewhere. The Riemann zeta function plays a pivotal role in analytic number theory, and has applications in physics, probability theory, and applied statistics. Leonhard Euler first introduced and studied the function over the reals in the first half of the eighteenth century.
Irrational numberIn mathematics, the irrational numbers (from in- prefix assimilated to ir- (negative prefix, privative) + rational) are all the real numbers that are not rational numbers. That is, irrational numbers cannot be expressed as the ratio of two integers. When the ratio of lengths of two line segments is an irrational number, the line segments are also described as being incommensurable, meaning that they share no "measure" in common, that is, there is no length ("the measure"), no matter how short, that could be used to express the lengths of both of the two given segments as integer multiples of itself.
Zipf's lawZipf's law (zɪf, ts͡ɪpf) is an empirical law that often holds, approximately, when a list of measured values is sorted in decreasing order. It states that the value of the nth entry is inversely proportional to n. The best known instance of Zipf's law applies to the frequency table of words in a text or corpus of natural language: Namely, it is usually found that the most common word occurs approximately twice as often as the next common one, three times as often as the third most common, and so on.
Divergent seriesIn mathematics, a divergent series is an infinite series that is not convergent, meaning that the infinite sequence of the partial sums of the series does not have a finite limit. If a series converges, the individual terms of the series must approach zero. Thus any series in which the individual terms do not approach zero diverges. However, convergence is a stronger condition: not all series whose terms approach zero converge. A counterexample is the harmonic series The divergence of the harmonic series was proven by the medieval mathematician Nicole Oresme.
FactorialIn mathematics, the factorial of a non-negative integer , denoted by , is the product of all positive integers less than or equal to . The factorial of also equals the product of with the next smaller factorial: For example, The value of 0! is 1, according to the convention for an empty product. Factorials have been discovered in several ancient cultures, notably in Indian mathematics in the canonical works of Jain literature, and by Jewish mystics in the Talmudic book Sefer Yetzirah.
Analytic number theoryIn mathematics, analytic number theory is a branch of number theory that uses methods from mathematical analysis to solve problems about the integers. It is often said to have begun with Peter Gustav Lejeune Dirichlet's 1837 introduction of Dirichlet L-functions to give the first proof of Dirichlet's theorem on arithmetic progressions. It is well known for its results on prime numbers (involving the Prime Number Theorem and Riemann zeta function) and additive number theory (such as the Goldbach conjecture and Waring's problem).
Integral test for convergenceIn mathematics, the integral test for convergence is a method used to test infinite series of monotonous terms for convergence. It was developed by Colin Maclaurin and Augustin-Louis Cauchy and is sometimes known as the Maclaurin–Cauchy test. Consider an integer N and a function f defined on the unbounded interval , on which it is monotone decreasing. Then the infinite series converges to a real number if and only if the improper integral is finite. In particular, if the integral diverges, then the series diverges as well.
Alternating seriesIn mathematics, an alternating series is an infinite series of the form or with an > 0 for all n. The signs of the general terms alternate between positive and negative. Like any series, an alternating series converges if and only if the associated sequence of partial sums converges. The geometric series 1/2 − 1/4 + 1/8 − 1/16 + ⋯ sums to 1/3. The alternating harmonic series has a finite sum but the harmonic series does not.
Taylor seriesIn mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor series are equal near this point. Taylor series are named after Brook Taylor, who introduced them in 1715. A Taylor series is also called a Maclaurin series when 0 is the point where the derivatives are considered, after Colin Maclaurin, who made extensive use of this special case of Taylor series in the mid-18th century.
Conditional convergenceIn mathematics, a series or integral is said to be conditionally convergent if it converges, but it does not converge absolutely. More precisely, a series of real numbers is said to converge conditionally if exists (as a finite real number, i.e. not or ), but A classic example is the alternating harmonic series given by which converges to , but is not absolutely convergent (see Harmonic series). Bernhard Riemann proved that a conditionally convergent series may be rearranged to converge to any value at all, including ∞ or −∞; see Riemann series theorem.
Sine waveA sine wave, sinusoidal wave, or sinusoid is a mathematical curve defined in terms of the sine trigonometric function, of which it is the graph. It is a type of continuous wave and also a smooth periodic function. It occurs often in mathematics, as well as in physics, engineering, signal processing and many other fields. Its most basic form as a function of time (t) is: where: A, amplitude, the peak deviation of the function from zero. f, ordinary frequency, the number of oscillations (cycles) that occur each second of time.
Digamma functionIn mathematics, the digamma function is defined as the logarithmic derivative of the gamma function: It is the first of the polygamma functions. This function is strictly increasing and strictly concave on , and it asymptotically behaves as for large arguments () in the sector with some infinitesimally small positive constant . The digamma function is often denoted as or Ϝ (the uppercase form of the archaic Greek consonant digamma meaning double-gamma).
1 + 1 + 1 + 1 + ⋯In mathematics, 1 + 1 + 1 + 1 + ⋯, also written \sum_{n=1}^{\infin} n^0, , or simply , is a divergent series, meaning that its sequence of partial sums does not converge to a limit in the real numbers. The sequence 1n can be thought of as a geometric series with the common ratio 1. Unlike other geometric series with rational ratio (except −1), it converges in neither the real numbers nor in the p-adic numbers for some p. In the context of the extended real number line since its sequence of partial sums increases monotonically without bound.