Telescoping series
In-game article clicks load inline without leaving the challenge.
In mathematics, a telescoping series is a series whose general term t n {\displaystyle t_{n}} is of the form t n = a n + 1 − a n {\displaystyle t_{n}=a_{n+1}-a_{n}}, i.e. the difference of two consecutive terms of a sequence ( a n ) {\displaystyle (a_{n})}. As a consequence the partial sums of the series only consists of two terms of ( a n ) {\displaystyle (a_{n})} after cancellation.
The cancellation technique, with part of each term cancelling with part of the next term, is known as the method of differences.
An early statement of the formula for the sum or partial sums of a telescoping series can be found in a 1644 work by Evangelista Torricelli, De dimensione parabolae.
Definition

Telescoping sums are finite sums in which pairs of consecutive terms partly cancel each other, leaving only parts of the initial and final terms. Let a n {\displaystyle a_{n}} be the elements of a sequence of numbers. Then ∑ n = 1 N ( a n − a n − 1 ) = a N − a 0 . {\displaystyle \sum _{n=1}^{N}\left(a_{n}-a_{n-1}\right)=a_{N}-a_{0}.} If a n {\displaystyle a_{n}} converges to a limit L {\displaystyle L}, the telescoping series gives: ∑ n = 1 ∞ ( a n − a n − 1 ) = L − a 0 . {\displaystyle \sum _{n=1}^{\infty }\left(a_{n}-a_{n-1}\right)=L-a_{0}.}
Every series is a telescoping series of its own partial sums.
Examples
- The product of a finite geometric series with initial term a {\displaystyle a} and common ratio r {\displaystyle r} by the factor ( 1 − r ) {\displaystyle (1-r)} yields a telescoping sum:( 1 − r ) ∑ k = 0 n a r k = ∑ k = 0 n ( a r k − a r k + 1 ) = a − a r n + 1 {\displaystyle (1-r)\sum _{k=0}^{n}ar^{k}=\sum _{k=0}^{n}\left(ar^{k}-ar^{k+1}\right)=a-ar^{n+1}} When | r | < 1 {\displaystyle |r|<1}, this allows for a direct calculation of its limit as n → ∞ {\displaystyle n\rightarrow \infty } and implies: ∑ k = 0 ∞ a r k = a 1 − r . {\displaystyle \sum _{k=0}^{\infty }ar^{k}={\frac {a}{1-r}}.}
- The series∑ n = 1 ∞ 1 n ( n + 1 ) {\displaystyle \sum _{n=1}^{\infty }{\frac {1}{n(n+1)}}}is the series of reciprocals of pronic numbers, and it is recognizable as a telescoping series once rewritten in partial fraction form ∑ n = 1 ∞ 1 n ( n + 1 ) = ∑ n = 1 ∞ ( 1 n − 1 n + 1 ) = lim N → ∞ ∑ n = 1 N ( 1 n − 1 n + 1 ) = lim N → ∞ [ ( 1 − 1 2 ) + ( 1 2 − 1 3 ) + ⋯ + ( 1 N − 1 N + 1 ) ] = lim N → ∞ [ 1 + ( − 1 2 + 1 2 ) + ( − 1 3 + 1 3 ) + ⋯ + ( − 1 N + 1 N ) − 1 N + 1 ] = lim N → ∞ [ 1 − 1 N + 1 ] = 1. {\displaystyle {\begin{aligned}\sum _{n=1}^{\infty }{\frac {1}{n(n+1)}}&{}=\sum _{n=1}^{\infty }\left({\frac {1}{n}}-{\frac {1}{n+1}}\right)\\{}&{}=\lim _{N\to \infty }\sum _{n=1}^{N}\left({\frac {1}{n}}-{\frac {1}{n+1}}\right)\\{}&{}=\lim _{N\to \infty }\left\lbrack {\left(1-{\frac {1}{2}}\right)+\left({\frac {1}{2}}-{\frac {1}{3}}\right)+\cdots +\left({\frac {1}{N}}-{\frac {1}{N+1}}\right)}\right\rbrack \\{}&{}=\lim _{N\to \infty }\left\lbrack {1+\left(-{\frac {1}{2}}+{\frac {1}{2}}\right)+\left(-{\frac {1}{3}}+{\frac {1}{3}}\right)+\cdots +\left(-{\frac {1}{N}}+{\frac {1}{N}}\right)-{\frac {1}{N+1}}}\right\rbrack \\{}&{}=\lim _{N\to \infty }\left\lbrack {1-{\frac {1}{N+1}}}\right\rbrack =1.\end{aligned}}}
- Let k be a positive integer. Then∑ n = 1 ∞ 1 n ( n + k ) = H k k {\displaystyle \sum _{n=1}^{\infty }{\frac {1}{n(n+k)}}={\frac {H_{k}}{k}}} where Hk is the kth harmonic number.
- Let k and m with k ≠ {\displaystyle \neq } m be positive integers. Then∑ n = 1 ∞ 1 ( n + k ) ( n + k + 1 ) … ( n + m − 1 ) ( n + m ) = 1 m − k ⋅ k ! m ! {\displaystyle \sum _{n=1}^{\infty }{\frac {1}{(n+k)(n+k+1)\dots (n+m-1)(n+m)}}={\frac {1}{m-k}}\cdot {\frac {k!}{m!}}} where ! {\displaystyle !} denotes the factorial operation.
- Many trigonometric functions also admit representation as differences, which may reveal telescopic canceling between the consecutive terms. Using the angle addition identity for a product of sines,∑ n = 1 N sin ( n ) = ∑ n = 1 N 1 2 csc ( 1 2 ) ( 2 sin ( 1 2 ) sin ( n ) ) = 1 2 csc ( 1 2 ) ∑ n = 1 N ( cos ( 2 n − 1 2 ) − cos ( 2 n + 1 2 ) ) = 1 2 csc ( 1 2 ) ( cos ( 1 2 ) − cos ( 2 N + 1 2 ) ) , {\displaystyle {\begin{aligned}\sum _{n=1}^{N}\sin \left(n\right)&{}=\sum _{n=1}^{N}{\frac {1}{2}}\csc \left({\frac {1}{2}}\right)\left(2\sin \left({\frac {1}{2}}\right)\sin \left(n\right)\right)\\&{}={\frac {1}{2}}\csc \left({\frac {1}{2}}\right)\sum _{n=1}^{N}\left(\cos \left({\frac {2n-1}{2}}\right)-\cos \left({\frac {2n+1}{2}}\right)\right)\\&{}={\frac {1}{2}}\csc \left({\frac {1}{2}}\right)\left(\cos \left({\frac {1}{2}}\right)-\cos \left({\frac {2N+1}{2}}\right)\right),\end{aligned}}} which does not converge as N → ∞ . {\textstyle N\rightarrow \infty .}
Applications
In probability theory, a Poisson process is a stochastic process of which the simplest case involves "occurrences" at random times, the waiting time until the next occurrence having a memoryless exponential distribution, and the number of "occurrences" in any time interval having a Poisson distribution whose expected value is proportional to the length of the time interval. Let Xt be the number of "occurrences" before time t, and let Tx be the waiting time until the xth "occurrence". We seek the probability density function of the random variable Tx. We use the probability mass function for the Poisson distribution, which tells us that
Pr ( X t = x ) = ( λ t ) x e − λ t x ! , {\displaystyle \Pr(X_{t}=x)={\frac {(\lambda t)^{x}e^{-\lambda t}}{x!}},}
where λ is the average number of occurrences in any time interval of length 1. Observe that the event {Xt ≥ x} is the same as the event {Tx ≤ t}, and thus they have the same probability. Intuitively, if something occurs at least x {\displaystyle x} times before time t {\displaystyle t}, we have to wait at most t {\displaystyle t} for the x t h {\displaystyle xth} occurrence. The density function we seek is therefore
f ( t ) = d d t Pr ( T x ≤ t ) = d d t Pr ( X t ≥ x ) = d d t ( 1 − Pr ( X t ≤ x − 1 ) ) = d d t ( 1 − ∑ u = 0 x − 1 Pr ( X t = u ) ) = d d t ( 1 − ∑ u = 0 x − 1 ( λ t ) u e − λ t u ! ) = λ e − λ t − e − λ t ∑ u = 1 x − 1 ( λ u t u − 1 ( u − 1 ) ! − λ u + 1 t u u ! ) {\displaystyle {\begin{aligned}f(t)&{}={\frac {d}{dt}}\Pr(T_{x}\leq t)={\frac {d}{dt}}\Pr(X_{t}\geq x)={\frac {d}{dt}}(1-\Pr(X_{t}\leq x-1))\\\\&{}={\frac {d}{dt}}\left(1-\sum _{u=0}^{x-1}\Pr(X_{t}=u)\right)={\frac {d}{dt}}\left(1-\sum _{u=0}^{x-1}{\frac {(\lambda t)^{u}e^{-\lambda t}}{u!}}\right)\\\\&{}=\lambda e^{-\lambda t}-e^{-\lambda t}\sum _{u=1}^{x-1}\left({\frac {\lambda ^{u}t^{u-1}}{(u-1)!}}-{\frac {\lambda ^{u+1}t^{u}}{u!}}\right)\end{aligned}}}
The sum telescopes, leaving
f ( t ) = λ x t x − 1 e − λ t ( x − 1 ) ! . {\displaystyle f(t)={\frac {\lambda ^{x}t^{x-1}e^{-\lambda t}}{(x-1)!}}.}
For other applications, see:
- Proof that the sum of the reciprocals of the primes diverges, where one of the proofs uses a telescoping sum;
- Fundamental theorem of calculus, a continuous analog of telescoping series;
- Order statistic, where a telescoping sum occurs in the derivation of a probability density function;
- Lefschetz fixed-point theorem, where a telescoping sum arises in algebraic topology;
- Homology theory, concept in algebraic topology;
- Eilenberg–Mazur swindle, where a telescoping sum of knots occurs;
- Faddeev–LeVerrier algorithm.
Related concepts
A telescoping product is a finite product (or the partial product of an infinite product) that can be canceled by the method of quotients to be eventually only a finite number of factors. It is the finite products in which consecutive terms cancel denominator with numerator, leaving only the initial and final terms. Let a n {\displaystyle a_{n}} be a sequence of numbers. Then, ∏ n = 1 N a n − 1 a n = a 0 a N . {\displaystyle \prod _{n=1}^{N}{\frac {a_{n-1}}{a_{n}}}={\frac {a_{0}}{a_{N}}}.} If a n {\displaystyle a_{n}} converges to 1, the resulting product gives: ∏ n = 1 ∞ a n − 1 a n = a 0 {\displaystyle \prod _{n=1}^{\infty }{\frac {a_{n-1}}{a_{n}}}=a_{0}}
For example, the infinite product ∏ n = 2 ∞ ( 1 − 1 n 2 ) {\displaystyle \prod _{n=2}^{\infty }\left(1-{\frac {1}{n^{2}}}\right)} simplifies as ∏ n = 2 ∞ ( 1 − 1 n 2 ) = ∏ n = 2 ∞ ( n − 1 ) ( n + 1 ) n 2 = lim N → ∞ ∏ n = 2 N n − 1 n × ∏ n = 2 N n + 1 n = lim N → ∞ [ 1 2 × 2 3 × 3 4 × ⋯ × N − 1 N ] × [ 3 2 × 4 3 × 5 4 × ⋯ × N N − 1 × N + 1 N ] = lim N → ∞ [ 1 2 ] × [ N + 1 N ] = 1 2 × lim N → ∞ [ N + 1 N ] = 1 2 . {\displaystyle {\begin{aligned}\prod _{n=2}^{\infty }\left(1-{\frac {1}{n^{2}}}\right)&=\prod _{n=2}^{\infty }{\frac {(n-1)(n+1)}{n^{2}}}\\&=\lim _{N\to \infty }\prod _{n=2}^{N}{\frac {n-1}{n}}\times \prod _{n=2}^{N}{\frac {n+1}{n}}\\&=\lim _{N\to \infty }\left\lbrack {{\frac {1}{2}}\times {\frac {2}{3}}\times {\frac {3}{4}}\times \cdots \times {\frac {N-1}{N}}}\right\rbrack \times \left\lbrack {{\frac {3}{2}}\times {\frac {4}{3}}\times {\frac {5}{4}}\times \cdots \times {\frac {N}{N-1}}\times {\frac {N+1}{N}}}\right\rbrack \\&=\lim _{N\to \infty }\left\lbrack {\frac {1}{2}}\right\rbrack \times \left\lbrack {\frac {N+1}{N}}\right\rbrack \\&={\frac {1}{2}}\times \lim _{N\to \infty }\left\lbrack {\frac {N+1}{N}}\right\rbrack \\&={\frac {1}{2}}.\end{aligned}}}