# 1.1: Basic properties

- Last updated

- Save as PDF

- Page ID
- 32278

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\id}{\mathrm{id}}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\kernel}{\mathrm{null}\,}\)

\( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\)

\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\)

\( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)

\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)

\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vectorC}[1]{\textbf{#1}} \)

\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)## Introduction

### About this book

This book is a one semester course in basic analysis. It started its life as my lecture notes for teaching Math 444 at the University of Illinois at Urbana-Champaign (UIUC) in Fall semester 2009. Later I added the metric space chapter to teach Math 521 at University of Wisconsin–Madison (UW). A prerequisite for this course is a basic proof course, using for example , , or .

It should be possible to use the book for both a basic course for students who do not necessarily wish to go to graduate school (such as UIUC 444), but also as a more advanced one-semester course that also covers topics such as metric spaces (such as UW 521). Here are my suggestions for what to cover in a semester course. For a slower course such as UIUC 444:

§0.3, §1.1–§1.4, §2.1–§2.5, §3.1–§3.4, §4.1–§4.2, §5.1–§5.3, §6.1–§6.3

For a more rigorous course covering metric spaces that runs quite a bit faster (such as UW 521):

§0.3, §1.1–§1.4, §2.1–§2.5, §3.1–§3.4, §4.1–§4.2, §5.1–§5.3, §6.1–§6.2, §7.1–§7.6

It should also be possible to run a faster course without metric spaces covering all sections of chapters 0 through 6. The approximate number of lectures given in the section notes through chapter 6 are a very rough estimate and were designed for the slower course. The first few chapters of the book can be used in an introductory proofs course as is for example done at Iowa State University Math 201, where this book is used in conjunction with Hammack’s Book of Proof .

The book normally used for the class at UIUC is Bartle and Sherbert, *Introduction to Real Analysis* third edition . The structure of the beginning of the book somewhat follows the standard syllabus of UIUC Math 444 and therefore has some similarities with . A major difference is that we define the Riemann integral using Darboux sums and not tagged partitions. The Darboux approach is far more appropriate for a course of this level.

Our approach allows us to fit a course such as UIUC 444 within a semester and still spend some extra time on the interchange of limits and end with Picard’s theorem on the existence and uniqueness of solutions of ordinary differential equations. This theorem is a wonderful example that uses many results proved in the book. For more advanced students, material may be covered faster so that we arrive at metric spaces and prove Picard’s theorem using the fixed point theorem as is usual.

Other excellent books exist. My favorite is Rudin’s excellent *Principles of Mathematical Analysis* or as it is commonly and lovingly called *baby Rudin* (to distinguish it from his other great analysis textbook). I took a lot of inspiration and ideas from Rudin. However, Rudin is a bit more advanced and ambitious than this present course. For those that wish to continue mathematics, Rudin is a fine investment. An inexpensive and somewhat simpler alternative to Rudin is Rosenlicht’s *Introduction to Analysis* . There is also the freely downloadable *Introduction to Real Analysis* by William Trench .

A note about the style of some of the proofs: Many proofs traditionally done by contradiction, I prefer to do by a direct proof or by contrapositive. While the book does include proofs by contradiction, I only do so when the contrapositive statement seemed too awkward, or when contradiction follows rather quickly. In my opinion, contradiction is more likely to get beginning students into trouble, as we are talking about objects that do not exist.

I try to avoid unnecessary formalism where it is unhelpful. Furthermore, the proofs and the language get slightly less formal as we progress through the book, as more and more details are left out to avoid clutter.

As a general rule, I use \(:=\) instead of \(=\) to define an object rather than to simply show equality. I use this symbol rather more liberally than is usual for emphasis. I use it even when the context is “local,” that is, I may simply define a function \(f(x) := x^2\) for a single exercise or example.

Finally, I would like to acknowledge Jana Maříková, Glen Pugh, Paul Vojta, Frank Beatrous, Sönmez Şahutoğlu, Jim Brandt, Kenji Kozai, and Arthur Busch, for teaching with the book and giving me lots of useful feedback. Frank Beatrous wrote the University of Pittsburgh version extensions, which served as inspiration for many of the recent additions. I would also like to thank Dan Stoneham, Jeremy Sutter, Eliya Gwetta, Daniel Pimentel-Alarcón, Steve Hoerning, Yi Zhang, Nicole Caviris, Kristopher Lee, Baoyue Bi, Hannah Lund, Trevor Mannella, Mitchel Meyer, Gregory Beauregard, Chase Meadors, Andreas Giannopoulos, an anonymous reader, and in general all the students in my classes for suggestions and finding errors and typos.

### About analysis

Analysis is the branch of mathematics that deals with inequalities and limits. The present course deals with the most basic concepts in analysis. The goal of the course is to acquaint the reader with rigorous proofs in analysis and also to set a firm foundation for calculus of one variable.

Calculus has prepared you, the student, for using mathematics without telling you why what you learned is true. To use, or teach, mathematics effectively, you cannot simply know *what* is true, you must know *why* it is true. This course shows you *why* calculus is true. It is here to give you a good understanding of the concept of a limit, the derivative, and the integral.

Let us use an analogy. An auto mechanic that has learned to change the oil, fix broken headlights, and charge the battery, will only be able to do those simple tasks. He will be unable to work independently to diagnose and fix problems. A high school teacher that does not understand the definition of the Riemann integral or the derivative may not be able to properly answer all the students’ questions. To this day I remember several nonsensical statements I heard from my calculus teacher in high school, who simply did not understand the concept of the limit, though he could “do” all problems in calculus.

We start with a discussion of the real number system, most importantly its completeness property, which is the basis for all that comes after. We then discuss the simplest form of a limit, the limit of a sequence. Afterwards, we study functions of one variable, continuity, and the derivative. Next, we define the Riemann integral and prove the fundamental theorem of calculus. We discuss sequences of functions and the interchange of limits. Finally, we give an introduction to metric spaces.

Let us give the most important difference between analysis and algebra. In algebra, we prove equalities directly; we prove that an object, a number perhaps, is equal to another object. In analysis, we usually prove inequalities. To illustrate the point, consider the following statement.

*Let \(x\) be a real number. If \(0 \leq x < \epsilon\) is true for all real numbers \(\epsilon > 0\), then \(x = 0\)*.

This statement is the general idea of what we do in analysis. If we wish to show that \(x = 0\), we show that \(0 \leq x < \epsilon\) for all positive \(\epsilon\).

The term *real analysis* is a little bit of a misnomer. I prefer to use simply *analysis*. The other type of analysis, *complex analysis*, really builds up on the present material, rather than being distinct. Furthermore, a more advanced course on real analysis would talk about complex numbers often. I suspect the nomenclature is historical baggage.

Let us get on with the show…

### Basic set theory

*Note: 1–3 lectures (some material can be skipped or covered lightly)*

Before we start talking about analysis we need to fix some language. Modern^{1} analysis uses the language of sets, and therefore that is where we start. We talk about sets in a rather informal way, using the so-called “naïve set theory.” Do not worry, that is what majority of mathematicians use, and it is hard to get into trouble.

We assume the reader has seen basic set theory and has had a course in basic proof writing. This section should be thought of as a refresher.

#### Sets

A *set* is a collection of objects called *elements* or *members*. A set with no objects is called the *empty set* and is denoted by \(\emptyset\) (or sometimes by \(\{ \}\)).

Think of a set as a club with a certain membership. For example, the students who play chess are members of the chess club. However, do not take the analogy too far. A set is only defined by the members that form the set; two sets that have the same members are the same set.

Most of the time we will consider sets of numbers. For example, the set \[S := \{ 0, 1, 2 \}\] is the set containing the three elements 0, 1, and 2. We write \[1 \in S\] to denote that the number 1 belongs to the set \(S\). That is, 1 is a member of \(S\). Similarly we write \[7 \notin S\] to denote that the number 7 is not in \(S\). That is, 7 is not a member of \(S\). The elements of all sets under consideration come from some set we call the *universe*. For simplicity, we often consider the universe to be the set that contains only the elements we are interested in. The universe is generally understood from context and is not explicitly mentioned. In this course, our universe will most often be the set of real numbers.

While the elements of a set are often numbers, other objects, such as other sets, can be elements of a set. A set may also contain some of the same elements as another set. For example, \[T := \{ 0, 2 \}\] contains the numbers 0 and 2. In this case all elements of \(T\) also belong to \(S\). We write \(T \subset S\). More formally we make the following definition.

- A set \(A\) is a
*subset*of a set \(B\) if \(x \in A\) implies \(x \in B\), and we write \(A \subset B\). That is, all members of \(A\) are also members of \(B\). - Two sets \(A\) and \(B\) are
*equal*if \(A \subset B\) and \(B \subset A\). We write \(A = B\). That is, \(A\) and \(B\) contain exactly the same elements. If it is not true that \(A\) and \(B\) are equal, then we write \(A \not= B\). - A set \(A\) is a
*proper subset*of \(B\) if \(A \subset B\) and \(A \not= B\). We write \(A \subsetneq B\).

For example, for \(S\) and \(T\) defined above \(T \subset S\), but \(T \not= S\). So \(T\) is a proper subset of \(S\). If \(A = B\), then \(A\) and \(B\) are simply two names for the same exact set. Let us mention the *set building notation*, \[\{ x \in A : P(x) \} .\] This notation refers to a subset of the set \(A\) containing all elements of \(A\) that satisfy the property \(P(x)\). The notation is sometimes abbreviated, \(A\) is not mentioned when understood from context. Furthermore, \(x \in A\) is sometimes replaced with a formula to make the notation easier to read.

The following are sets including the standard notations.

- The set of
*natural numbers*, \({\mathbb{N}}:= \{ 1, 2, 3, \ldots \}\). - The set of
*integers*, \({\mathbb{Z}}:= \{ 0, -1, 1, -2, 2, \ldots \}\). - The set of
*rational numbers*, \({\mathbb{Q}}:= \{ \frac{m}{n} : m, n \in {\mathbb{Z}} \text{ and } n \not= 0 \}\). - The set of even natural numbers, \(\{ 2m : m \in {\mathbb{N}}\}\).
- The set of real numbers, \({\mathbb{R}}\).

Note that \({\mathbb{N}}\subset {\mathbb{Z}}\subset {\mathbb{Q}}\subset {\mathbb{R}}\).

There are many operations we want to do with sets.

- A
*union*of two sets \(A\) and \(B\) is defined as \[A \cup B := \{ x : x \in A \text{ or } x \in B \} .\] - An
*intersection*of two sets \(A\) and \(B\) is defined as \[A \cap B := \{ x : x \in A \text{ and } x \in B \} .\] - A
*complement of \(B\) relative to \(A\)*(or*set-theoretic difference*of \(A\) and \(B\)) is defined as \[A \setminus B := \{ x : x \in A \text{ and } x \notin B \} .\] - We say
*complement*of \(B\) and write \(B^c\) instead of \(A \setminus B\) if the set \(A\) is either the entire universe or is the obvious set containing \(B\), and is understood from context. - We say sets \(A\) and \(B\) are
*disjoint*if \(A \cap B = \emptyset\).

The notation \(B^c\) may be a little vague at this point. If the set \(B\) is a subset of the real numbers \({\mathbb{R}}\), then \(B^c\) means \({\mathbb{R}}\setminus B\). If \(B\) is naturally a subset of the natural numbers, then \(B^c\) is \({\mathbb{N}}\setminus B\). If ambiguity would ever arise, we will use the set difference notation \(A \setminus B\).

We illustrate the operations on the *Venn diagrams* in . Let us now establish one of most basic theorems about sets and logic.

Let \(A, B, C\) be sets. Then \[\begin{aligned} {(B \cup C)}^c &= B^c \cap C^c , \\ {(B \cap C)}^c &= B^c \cup C^c ,\end{aligned}\] or, more generally, \[\begin{aligned} A \setminus (B \cup C) &= (A \setminus B) \cap (A \setminus C) , \\ A \setminus (B \cap C) &= (A \setminus B) \cup (A \setminus C) .\end{aligned}\]

The first statement is proved by the second statement if we assume the set \(A\) is our “universe.”

Let us prove \(A \setminus (B \cup C) = (A \setminus B) \cap (A \setminus C)\). Remember the definition of equality of sets. First, we must show that if \(x \in A \setminus (B \cup C)\), then \(x \in (A \setminus B) \cap (A \setminus C)\). Second, we must also show that if \(x \in (A \setminus B) \cap (A \setminus C)\), then \(x \in A \setminus (B \cup C)\).

So let us assume \(x \in A \setminus (B \cup C)\). Then \(x\) is in \(A\), but not in \(B\) nor \(C\). Hence \(x\) is in \(A\) and not in \(B\), that is, \(x \in A \setminus B\). Similarly \(x \in A \setminus C\). Thus \(x \in (A \setminus B) \cap (A \setminus C)\).

On the other hand suppose \(x \in (A \setminus B) \cap (A \setminus C)\). In particular \(x \in (A \setminus B)\), so \(x \in A\) and \(x \notin B\). Also as \(x \in (A \setminus C)\), then \(x \notin C\). Hence \(x \in A \setminus (B \cup C)\).

The proof of the other equality is left as an exercise.

We will also need to intersect or union several sets at once. If there are only finitely many, then we simply apply the union or intersection operation several times. However, suppose we have an infinite collection of sets (a set of sets) \(\{ A_1, A_2, A_3, \ldots \}\). We define \[\begin{aligned} & \bigcup_{n=1}^\infty A_n := \{ x : x \in A_n \text{ for some $n \in {\mathbb{N}}$} \} , \\ & \bigcap_{n=1}^\infty A_n := \{ x : x \in A_n \text{ for all $n \in {\mathbb{N}}$} \} .\end{aligned}\]

We can also have sets indexed by two integers. For example, we can have the set of sets \(\{ A_{1,1}, A_{1,2}, A_{2,1}, A_{1,3}, A_{2,2}, A_{3,1}, \ldots \}\). Then we write \[\bigcup_{n=1}^\infty \bigcup_{m=1}^\infty A_{n,m} = \bigcup_{n=1}^\infty \left( \bigcup_{m=1}^\infty A_{n,m} \right) .\] And similarly with intersections.

It is not hard to see that we can take the unions in any order. However, switching the order of unions and intersections is not generally permitted without proof. For example: \[\bigcup_{n=1}^\infty \bigcap_{m=1}^\infty \{ k \in {\mathbb{N}}: mk < n \} = \bigcup_{n=1}^\infty \emptyset = \emptyset .\] However, \[\bigcap_{m=1}^\infty \bigcup_{n=1}^\infty \{ k \in {\mathbb{N}}: mk < n \} = \bigcap_{m=1}^\infty {\mathbb{N}} = {\mathbb{N}}.\]

Sometimes, the index set is not the natural numbers. In this case we need a more general notation. Suppose \(I\) is some set and for each \(\iota \in I\), we have a set \(A_\iota\). Then we define \[\bigcup_{\iota \in I} A_\iota := \{ x : x \in A_\iota \text{ for some $\iota \in I$} \} \qquad \bigcap_{\iota \in I} A_\iota := \{ x : x \in A_\iota \text{ for all $\iota \in I$} \} .\]

#### Induction

When a statement includes an arbitrary natural number, a common method of proof is the principle of induction. We start with the set of natural numbers \({\mathbb{N}}= \{ 1,2,3,\ldots \}\), and we give them their natural ordering, that is, \(1 < 2 < 3 < 4 < \cdots\). By \(S \subset {\mathbb{N}}\) having a *least element*, we mean that there exists an \(x \in S\), such that for every \(y \in S\), we have \(x \leq y\).

The natural numbers \({\mathbb{N}}\) ordered in the natural way possess the so-called *well ordering property*. We take this property as an axiom; we simply assume it is true.

Every nonempty subset of \({\mathbb{N}}\) has a least (smallest) element.

The *principle of induction* is the following theorem, which is equivalent to the well ordering property of the natural numbers.

[induction:thm] Let \(P(n)\) be a statement depending on a natural number \(n\). Suppose that

*(basis statement)*\(P(1)\) is true,*(induction step)*if \(P(n)\) is true, then \(P(n+1)\) is true.

Then \(P(n)\) is true for all \(n \in {\mathbb{N}}\).

Suppose \(S\) is the set of natural numbers \(m\) for which \(P(m)\) is not true. Suppose \(S\) is nonempty. Then \(S\) has a least element by the well ordering property. Let us call \(m\) the least element of \(S\). We know \(1 \notin S\) by assumption. Therefore \(m > 1\) and \(m-1\) is a natural number as well. Since \(m\) was the least element of \(S\), we know that \(P(m-1)\) is true. But by the induction step we see that \(P(m-1+1) = P(m)\) is true, contradicting the statement that \(m \in S\). Therefore \(S\) is empty and \(P(n)\) is true for all \(n \in {\mathbb{N}}\).

Sometimes it is convenient to start at a different number than 1, but all that changes is the labeling. The assumption that \(P(n)\) is true in “if \(P(n)\) is true, then \(P(n+1)\) is true” is usually called the *induction hypothesis*.

Let us prove that for all \(n \in {\mathbb{N}}\), \[2^{n-1} \leq n! .\] We let \(P(n)\) be the statement that \(2^{n-1} \leq n!\) is true. By plugging in \(n=1\), we see that \(P(1)\) is true.

Suppose \(P(n)\) is true. That is, suppose \(2^{n-1} \leq n!\) holds. Multiply both sides by 2 to obtain \[2^n \leq 2(n!) .\] As \(2 \leq (n+1)\) when \(n \in {\mathbb{N}}\), we have \(2(n!) \leq (n+1)(n!) = (n+1)!\). That is, \[2^n \leq 2(n!) \leq (n+1)!,\] and hence \(P(n+1)\) is true. By the principle of induction, we see that \(P(n)\) is true for all \(n\), and hence \(2^{n-1} \leq n!\) is true for all \(n \in {\mathbb{N}}\).

We claim that for all \(c \not= 1\), \[1 + c + c^2 + \cdots + c^n = \frac{1-c^{n+1}}{1-c} .\]

Proof: It is easy to check that the equation holds with \(n=1\). Suppose it is true for \(n\). Then \[\begin{split} 1 + c + c^2 + \cdots + c^n + c^{n+1} & = ( 1 + c + c^2 + \cdots + c^n ) + c^{n+1} \\ & = \frac{1-c^{n+1}}{1-c} + c^{n+1} \\ & = \frac{1-c^{n+1} + (1-c)c^{n+1}}{1-c} \\ & = \frac{1-c^{n+2}}{1-c} . \end{split}\]

There is an equivalent principle called strong induction. The proof that strong induction is equivalent to induction is left as an exercise.

Let \(P(n)\) be a statement depending on a natural number \(n\). Suppose that

*(basis statement)*\(P(1)\) is true,*(induction step)*if \(P(k)\) is true for all \(k = 1,2,\ldots,n\), then \(P(n+1)\) is true.

Then \(P(n)\) is true for all \(n \in {\mathbb{N}}\).

#### Functions

Informally, a *set-theoretic function* \(f\) taking a set \(A\) to a set \(B\) is a mapping that to each \(x \in A\) assigns a unique \(y \in B\). We write \(f \colon A \to B\). For example, we define a function \(f \colon S \to T\) taking \(S = \{ 0, 1, 2 \}\) to \(T = \{ 0, 2 \}\) by assigning \(f(0) := 2\), \(f(1) := 2\), and \(f(2) := 0\). That is, a function \(f \colon A \to B\) is a black box, into which we stick an element of \(A\) and the function spits out an element of \(B\). Sometimes \(f\) is called a *mapping* and we say \(f\) *maps \(A\) to \(B\)*.

Often, functions are defined by some sort of formula, however, you should really think of a function as just a very big table of values. The subtle issue here is that a single function can have several different formulas, all giving the same function. Also, for many functions, there is no formula that expresses its values.

To define a function rigorously first let us define the Cartesian product.

Let \(A\) and \(B\) be sets. The *Cartesian product* is the set of tuples defined as \[A \times B := \{ (x,y) : x \in A, y \in B \} .\]

For example, the set \([0,1] \times [0,1]\) is a set in the plane bounded by a square with vertices \((0,0)\), \((0,1)\), \((1,0)\), and \((1,1)\). When \(A\) and \(B\) are the same set we sometimes use a superscript 2 to denote such a product. For example \([0,1]^2 = [0,1] \times [0,1]\), or \({\mathbb{R}}^2 = {\mathbb{R}}\times {\mathbb{R}}\) (the Cartesian plane).

A *function* \(f \colon A \to B\) is a subset \(f\) of \(A \times B\) such that for each \(x \in A\), there is a unique \((x,y) \in f\). We then write \(f(x) = y\). Sometimes the set \(f\) is called the *graph* of the function rather than the function itself.

The set \(A\) is called the *domain* of \(f\) (and sometimes confusingly denoted \(D(f)\)). The set \[R(f) := \{ y \in B : \text{there exists an $x$ such that %$(x,y) \in f$ $f(x)=y$ } \}\] is called the *range* of \(f\).

Note that \(R(f)\) can possibly be a proper subset of \(B\), while the domain of \(f\) is always equal to \(A\). We usually assume that the domain of \(f\) is nonempty.

From calculus, you are most familiar with functions taking real numbers to real numbers. However, you saw some other types of functions as well. For example, the derivative is a function mapping the set of differentiable functions to the set of all functions. Another example is the Laplace transform, which also takes functions to functions. Yet another example is the function that takes a continuous function \(g\) defined on the interval \([0,1]\) and returns the number \(\int_0^1 g(x) dx\).

Let \(f \colon A \to B\) be a function, and \(C \subset A\). Define the *image* (or *direct image*) of \(C\) as \[f(C) := \{ f(x) \in B : x \in C \} .\] Let \(D \subset B\). Define the *inverse image* as \[f^{-1}(D) := \{ x \in A : f(x) \in D \} .\]

Define the function \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) by \(f(x) := \sin(\pi x)\). Then \(f([0,\nicefrac{1}{2}]) = [0,1]\), \(f^{-1}(\{0\}) = {\mathbb{Z}}\), etc….

[st:propinv] Let \(f \colon A \to B\). Let \(C, D\) be subsets of \(B\). Then \[\begin{aligned} & f^{-1}( C \cup D) = f^{-1} (C) \cup f^{-1} (D) , \\ & f^{-1}( C \cap D) = f^{-1} (C) \cap f^{-1} (D) , \\ & f^{-1}( C^c) = {\left( f^{-1} (C) \right)}^c .\end{aligned}\]

Read the last line as \(f^{-1}( B \setminus C) = A \setminus f^{-1} (C)\).

Let us start with the union. Suppose \(x \in f^{-1}( C \cup D)\). That means \(x\) maps to \(C\) or \(D\). Thus \(f^{-1}( C \cup D) \subset f^{-1} (C) \cup f^{-1} (D)\). Conversely if \(x \in f^{-1}(C)\), then \(x \in f^{-1}(C \cup D)\). Similarly for \(x \in f^{-1}(D)\). Hence \(f^{-1}( C \cup D) \supset f^{-1} (C) \cup f^{-1} (D)\), and we have equality.

The rest of the proof is left as an exercise.

The proposition does not hold for direct images. We do have the following weaker result.

[st:propfor] Let \(f \colon A \to B\). Let \(C, D\) be subsets of \(A\). Then \[\begin{aligned} & f( C \cup D) = f (C) \cup f (D) , \\ & f( C \cap D) \subset f (C) \cap f (D) .\end{aligned}\]

The proof is left as an exercise.

Let \(f \colon A \to B\) be a function. The function \(f\) is said to be *injective* or *one-to-one* if \(f(x_1) = f(x_2)\) implies \(x_1 = x_2\). In other words, for all \(y \in B\) the set \(f^{-1}(\{y\})\) is empty or consists of a single element. We call such an \(f\) an *injection*.

The function \(f\) is said to be *surjective* or *onto* if \(f(A) = B\). We call such an \(f\) a *surjection*.

A function \(f\) that is both an injection and a surjection is said to be *bijective*, and we say \(f\) is a *bijection*.

When \(f \colon A \to B\) is a bijection, then \(f^{-1}(\{y\})\) is always a unique element of \(A\), and we can consider \(f^{-1}\) as a function \(f^{-1} \colon B \to A\). In this case, we call \(f^{-1}\) the *inverse function* of \(f\). For example, for the bijection \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) defined by \(f(x) := x^3\) we have \(f^{-1}(x) = \sqrt[3]{x}\).

A final piece of notation for functions that we need is the *composition of functions*.

Let \(f \colon A \to B\), \(g \colon B \to C\). The function \(g \circ f \colon A \to C\) is defined as \[(g \circ f)(x) := g\bigl(f(x)\bigr) .\]

#### Cardinality

A subtle issue in set theory and one generating a considerable amount of confusion among students is that of cardinality, or “size” of sets. The concept of cardinality is important in modern mathematics in general and in analysis in particular. In this section, we will see the first really unexpected theorem.

Let \(A\) and \(B\) be sets. We say \(A\) and \(B\) have the same *cardinality* when there exists a bijection \(f \colon A \to B\). We denote by \(\left\lvert {A} \right\rvert\) the equivalence class of all sets with the same cardinality as \(A\) and we simply call \(\left\lvert {A} \right\rvert\) the cardinality of \(A\).

Note that \(A\) has the same cardinality as the empty set if and only if \(A\) itself is the empty set. We then write \(\left\lvert {A} \right\rvert := 0\).

Suppose \(A\) has the same cardinality as \(\{ 1,2,3,\ldots,n \}\) for some \(n \in {\mathbb{N}}\). We then write \(\left\lvert {A} \right\rvert := n\), and we say \(A\) is *finite*. When \(A\) is the empty set, we also call \(A\) finite.

We say \(A\) is *infinite* or “of infinite cardinality” if \(A\) is not finite.

That the notation \(\left\lvert {A} \right\rvert = n\) is justified we leave as an exercise. That is, for each nonempty finite set \(A\), there exists a unique natural number \(n\) such that there exists a bijection from \(A\) to \(\{ 1,2,3,\ldots,n \}\).

We can order sets by size.

[def:comparecards] We write \[\left\lvert {A} \right\rvert \leq \left\lvert {B} \right\rvert\] if there exists an injection from \(A\) to \(B\). We write \(\left\lvert {A} \right\rvert = \left\lvert {B} \right\rvert\) if \(A\) and \(B\) have the same cardinality. We write \(\left\lvert {A} \right\rvert < \left\lvert {B} \right\rvert\) if \(\left\lvert {A} \right\rvert \leq \left\lvert {B} \right\rvert\), but \(A\) and \(B\) do not have the same cardinality.

We state without proof that \(\left\lvert {A} \right\rvert = \left\lvert {B} \right\rvert\) have the same cardinality if and only if \(\left\lvert {A} \right\rvert \leq \left\lvert {B} \right\rvert\) and \(\left\lvert {B} \right\rvert \leq \left\lvert {A} \right\rvert\). This is the so-called Cantor-Bernstein-Schroeder theorem. Furthermore, if \(A\) and \(B\) are any two sets, we can always write \(\left\lvert {A} \right\rvert \leq \left\lvert {B} \right\rvert\) or \(\left\lvert {B} \right\rvert \leq \left\lvert {A} \right\rvert\). The issues surrounding this last statement are very subtle. As we do not require either of these two statements, we omit proofs.

The truly interesting cases of cardinality are infinite sets. We start with the following definition.

If \(\left\lvert {A} \right\rvert = \left\lvert

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[1]/div[3]/div[4]/p[11]/span, line 1, column 1 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

*countably infinite*. If \(A\) is finite or countably infinite, then we say \(A\) is

*countable*. If \(A\) is not countable, then \(A\) is said to be

*uncountable*.

The cardinality of \({\mathbb{N}}\) is usually denoted as \(\aleph_0\) (read as aleph-naught)^{2}.

The set of even natural numbers has the same cardinality as \({\mathbb{N}}\). Proof: Given an even natural number, write it as \(2n\) for some \(n \in {\mathbb{N}}\). Then create a bijection taking \(2n\) to \(n\).

In fact, let us mention without proof the following characterization of infinite sets: *A set is infinite if and only if it is in one-to-one correspondence with a proper subset of itself*.

\({\mathbb{N}}\times {\mathbb{N}}\) is a countably infinite set. Proof: Arrange the elements of \({\mathbb{N}}\times {\mathbb{N}}\) as follows \((1,1)\), \((1,2)\), \((2,1)\), \((1,3)\), \((2,2)\), \((3,1)\), …. That is, always write down first all the elements whose two entries sum to \(k\), then write down all the elements whose entries sum to \(k+1\) and so on. Then define a bijection with \({\mathbb{N}}\) by letting 1 go to \((1,1)\), 2 go to \((1,2)\) and so on.

The set of rational numbers is countable. Proof: (informal) Follow the same procedure as in the previous example, writing \(\nicefrac{1}{1}\), \(\nicefrac{1}{2}\), \(\nicefrac{2}{1}\), etc…. However, leave out any fraction (such as \(\nicefrac{2}{2}\)) that has already appeared.

For completeness we mention the following statement. *If \(A \subset B\) and \(B\) is countable, then \(A\) is countable. Similarly if \(A\) is uncountable, then \(B\) is uncountable*. As we will not need this statement in the sequel, and as the proof requires the Cantor-Bernstein-Schroeder theorem mentioned above, we will not give it here.

We give the first truly striking result. First, we need a notation for the set of all subsets of a set.

If \(A\) is a set, we define the *power set* of \(A\), denoted by \({\mathcal{P}}(A)\), to be the set of all subsets of \(A\).

For example, if \(A := \{ 1,2\}\), then \({\mathcal{P}}(A) = \{ \emptyset, \{ 1 \}, \{ 2 \}, \{ 1, 2 \} \}\). For a finite set \(A\) of cardinality \(n\), the cardinality of \({\mathcal{P}}(A)\) is \(2^n\). This fact is left as an exercise. Hence, for finite sets the cardinality of \({\mathcal{P}}(A)\) is strictly larger than the cardinality of \(A\). What is an unexpected and striking fact is that this statement is still true for infinite sets.

\(\left\lvert {A} \right\rvert < \left\lvert

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[1]/div[3]/div[4]/p[21]/span, line 1, column 1 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

There exists an injection \(f \colon A \to {\mathcal{P}}(A)\). For any \(x \in A\), define \(f(x) := \{ x \}\). Therefore \(\left\lvert {A} \right\rvert \leq \left\lvert {{\mathcal{P}}(A)} \right\rvert\).

To finish the proof, we must show that no function \(f \colon A \to {\mathcal{P}}(A)\) is a surjection. Suppose \(f \colon A \to {\mathcal{P}}(A)\) is a function. So for \(x \in A\), \(f(x)\) is a subset of \(A\). Define the set \[B := \{ x \in A : x \notin f(x) \} .\] We claim that \(B\) is not in the range of \(f\) and hence \(f\) is not a surjection. Suppose there exists an \(x_0\) such that \(f(x_0) = B\). Either \(x_0 \in B\) or \(x_0 \notin B\). If \(x_0 \in B\), then \(x_0 \notin f(x_0) = B\), which is a contradiction. If \(x_0 \notin B\), then \(x_0 \in f(x_0) = B\), which is again a contradiction. Thus such an \(x_0\) does not exist. Therefore, \(B\) is not in the range of \(f\), and \(f\) is not a surjection. As \(f\) was an arbitrary function, no surjection exists.

One particular consequence of this theorem is that there do exist uncountable sets, as \({\mathcal{P}}({\mathbb{N}})\) must be uncountable. A related fact is that the set of real numbers (which we study in the next chapter) is uncountable. The existence of uncountable sets may seem unintuitive, and the theorem caused quite a controversy at the time it was announced. The theorem not only says that uncountable sets exist, but that there in fact exist progressively larger and larger infinite sets \({\mathbb{N}}\), \({\mathcal{P}}({\mathbb{N}})\), \({\mathcal{P}}({\mathcal{P}}({\mathbb{N}}))\), \({\mathcal{P}}({\mathcal{P}}({\mathcal{P}}({\mathbb{N}})))\), etc….

## Exercises

Show \(A \setminus (B \cap C) = (A \setminus B) \cup (A \setminus C)\).

Prove that the principle of strong induction is equivalent to the standard induction.

Finish the proof of .

- Prove .
- Find an example for which equality of sets in \(f( C \cap D) \subset f (C) \cap f (D)\) fails. That is, find an \(f\), \(A\), \(B\), \(C\), and \(D\) such that \(f( C \cap D)\) is a proper subset of \(f(C) \cap f(D)\).

Prove that if \(A\) is finite, then there exists a unique number \(n\) such that there exists a bijection between \(A\) and \(\{ 1, 2, 3, \ldots, n \}\). In other words, the notation \(\left\lvert {A} \right\rvert := n\) is justified. Hint: Show that if \(n > m\), then there is no injection from \(\{ 1, 2, 3, \ldots, n \}\) to \(\{ 1, 2, 3, \ldots, m \}\).

Prove

- \(A \cap (B \cup C) = (A \cap B) \cup (A \cap C)\)
- \(A \cup (B \cap C) = (A \cup B) \cap (A \cup C)\)

Let \(A \Delta B\) denote the *symmetric difference*, that is, the set of all elements that belong to either \(A\) or \(B\), but not to both \(A\) and \(B\).

- Draw a Venn diagram for \(A \Delta B\).
- Show \(A \Delta B = (A \setminus B) \cup (B \setminus A)\).
- Show \(A \Delta B = (A \cup B) \setminus ( A \cap B)\).

For each \(n \in {\mathbb{N}}\), let \(A_n := \{ (n+1)k : k \in {\mathbb{N}}\}\).

- Find \(A_1 \cap A_2\).
- Find \(\bigcup_{n=1}^\infty A_n\).
- Find \(\bigcap_{n=1}^\infty A_n\).

Determine \({\mathcal{P}}(S)\) (the power set) for each of the following:

- \(S = \emptyset\),
- \(S = \{1\}\),
- \(S = \{1,2\}\),
- \(S = \{1,2,3,4\}\).

Let \(f \colon A \to B\) and \(g \colon B \to C\) be functions.

- Prove that if \(g \circ f\) is injective, then \(f\) is injective.
- Prove that if \(g \circ f\) is surjective, then \(g\) is surjective.
- Find an explicit example where \(g \circ f\) is bijective, but neither \(f\) nor \(g\) are bijective.

Prove that \(n < 2^n\) by induction.

Show that for a finite set \(A\) of cardinality \(n\), the cardinality of \({\mathcal{P}}(A)\) is \(2^n\).

Prove \(\frac{1}{1\cdot 2} + \frac{1}{2\cdot 3} + \cdots + \frac{1}{n(n+1)} = \frac{n}{n+1}\) for all \(n \in {\mathbb{N}}\).

Prove \(1^3 + 2^3 + \cdots + n^3 = {\left( \frac{n(n+1)}{2} \right)}^2\) for all \(n \in {\mathbb{N}}\).

Prove that \(n^3 + 5n\) is divisible by \(6\) for all \(n \in {\mathbb{N}}\).

Find the smallest \(n \in {\mathbb{N}}\) such that \(2{(n+5)}^2 < n^3\) and call it \(n_0\). Show that \(2{(n+5)}^2 < n^3\) for all \(n \geq n_0\).

Find all \(n \in {\mathbb{N}}\) such that \(n^2 < 2^n\).

Finish the proof that the is equivalent to the well ordering property of \({\mathbb{N}}\). That is, prove the well ordering property for \({\mathbb{N}}\) using the principle of induction.

Give an example of a countable collection of finite sets \(A_1, A_2, \ldots\), whose union is not a finite set.

Give an example of a countable collection of infinite sets \(A_1, A_2, \ldots\), with \(A_j \cap A_k\) being infinite for all \(j\) and \(k\), such that \(\bigcap_{j=1}^\infty A_j\) is nonempty and finite.

## Real Numbers

### Basic properties

*Note: 1.5 lectures*

The main object we work with in analysis is the set of real numbers. As this set is so fundamental, often much time is spent on formally constructing the set of real numbers. However, we take an easier approach here and just assume that a set with the correct properties exists. We need to start with the definitions of those properties.

An *ordered set* is a set \(S\), together with a relation \(<\) such that

- For any \(x, y \in S\), exactly one of \(x < y\), \(x=y\), or \(y < x\) holds.
- If \(x < y\) and \(y < z\), then \(x < z\).

We write \(x \leq y\) if \(x < y\) or \(x=y\). We define \(>\) and \(\geq\) in the obvious way.

For example, the set of rational numbers \({\mathbb{Q}}\) is an ordered set by letting \(x < y\) if and only if \(y-x\) is a positive rational number, that is if \(y-x = \nicefrac{p}{q}\) where \(p,q \in {\mathbb{N}}\). Similarly, \({\mathbb{N}}\) and \({\mathbb{Z}}\) are also ordered sets.

There are other ordered sets than sets of numbers. For example, the set of countries can be ordered by landmass, so for example India \(>\) Lichtenstein. Any time you sort a set in some way, you are making an ordered set. A typical ordered set that you have used since primary school is the dictionary. It is the ordered set of words where the order is the so-called lexicographic ordering. Such ordered sets appear often for example in computer science. In this class we will mostly be interested in ordered set of numbers however.

Let \(E \subset S\), where \(S\) is an ordered set.

- If there exists a \(b \in S\) such that \(x \leq b\) for all \(x \in E\), then we say \(E\) is
*bounded above*and \(b\) is an*upper bound*of \(E\). - If there exists a \(b \in S\) such that \(x \geq b\) for all \(x \in E\), then we say \(E\) is
*bounded below*and \(b\) is a*lower bound*of \(E\). - If there exists an upper bound \(b_0\) of \(E\) such that whenever \(b\) is any upper bound for \(E\) we have \(b_0 \leq b\), then \(b_0\) is called the
*least upper bound*or the*supremum*of \(E\). We write \[\sup\, E := b_0 .\] - Similarly, if there exists a lower bound \(b_0\) of \(E\) such that whenever \(b\) is any lower bound for \(E\) we have \(b_0 \geq b\), then \(b_0\) is called the
*greatest lower bound*or the*infimum*of \(E\). We write \[\inf\, E := b_0 .\]

When a set \(E\) is both bounded above and bounded below, we say simply that \(E\) is *bounded*.

A supremum or infimum for \(E\) (even if they exist) need not be in \(E\). For example, the set \(E := \{ x \in {\mathbb{Q}}: x < 1 \}\) has a least upper bound of 1, but 1 is not in the set \(E\) itself. On the other hand, if we take \(G := \{ x \in {\mathbb{Q}}: x \leq 1 \}\), then the least upper bound of \(G\) is clearly also 1, and in this case \(1 \in G\). On the other hand, the set \(P := \{ x \in {\mathbb{Q}}: x \geq 0 \}\) has no upper bound (why?) and therefore it can have no least upper bound. On the other hand 0 is the greatest lower bound of \(P\).

[defn:lub] An ordered set \(S\) has the *least-upper-bound property* if every nonempty subset \(E \subset S\) that is bounded above has a least upper bound, that is \(\sup\, E\) exists in \(S\).

The *least-upper-bound property* is sometimes called the *completeness property* or the *Dedekind completeness property*.

The set \({\mathbb{Q}}\) of rational numbers does not have the least-upper-bound property. The subset \(\{ x \in {\mathbb{Q}}: x^2 < 2 \}\) does not have a supremum in \({\mathbb{Q}}\). The obvious supremum \(\sqrt{2}\) is not rational. Suppose \(x \in {\mathbb{Q}}\) such that \(x^2 = 2\). Write \(x=\nicefrac{m}{n}\) in lowest terms. So \({(\nicefrac{m}{n})}^2 = 2\) or \(m^2 = 2n^2\). Hence \(m^2\) is divisible by 2 and so \(m\) is divisible by 2. Write \(m = 2k\) and so \({(2k)}^2 = 2n^2\). Divide by 2 and note that \(2k^2 = n^2\), and hence \(n\) is divisible by 2. But that is a contradiction as \(\nicefrac{m}{n}\) was in lowest terms.

That \({\mathbb{Q}}\) does not have the least-upper-bound property is one of the most important reasons why we work with \({\mathbb{R}}\) in analysis. The set \({\mathbb{Q}}\) is just fine for algebraists. But analysts require the least-upper-bound property to do any work. We also require our real numbers to have many algebraic properties. In particular, we require that they are a field.

A set \(F\) is called a *field* if it has two operations defined on it, addition \(x+y\) and multiplication \(xy\), and if it satisfies the following axioms.

- If \(x \in F\) and \(y \in F\), then \(x+y \in F\).
*(commutativity of addition)*\(x+y = y+x\) for all \(x,y \in F\).*(associativity of addition)*\((x+y)+z = x+(y+z)\) for all \(x,y,z \in F\).- There exists an element \(0 \in F\) such that \(0+x = x\) for all \(x \in F\).
- For every element \(x\in F\) there exists an element \(-x \in F\) such that \(x + (-x) = 0\).

- If \(x \in F\) and \(y \in F\), then \(xy \in F\).
*(commutativity of multiplication)*\(xy = yx\) for all \(x,y \in F\).*(associativity of multiplication)*\((xy)z = x(yz)\) for all \(x,y,z \in F\).- There exists an element \(1 \in F\) (and \(1 \not= 0\)) such that \(1x = x\) for all \(x \in F\).
- For every \(x\in F\) such that \(x \not= 0\) there exists an element \(\nicefrac{1}{x} \in F\) such that \(x(\nicefrac{1}{x}) = 1\).
*(distributive law)*\(x(y+z) = xy+xz\) for all \(x,y,z \in F\).

The set \({\mathbb{Q}}\) of rational numbers is a field. On the other hand \({\mathbb{Z}}\) is not a field, as it does not contain multiplicative inverses. For example, there is no \(x \in {\mathbb{Z}}\) such that \(2x = 1\), so (M5) is not satisfied. You can check that (M5) is the only property that fails^{3}.

We will assume the basic facts about fields that are easily proved from the axioms. For example, \(0x = 0\) is easily proved by noting that \(xx = (0+x)x = 0x+xx\), using (A4), (D), and (M2). Then using (A5) on \(xx\), along with (A2), (A3), and (A4), we obtain \(0 = 0x\).

A field \(F\) is said to be an *ordered field* if \(F\) is also an ordered set such that:

- [defn:ordfield:i] For \(x,y,z \in F\), \(x < y\) implies \(x+z < y+z\).
- [defn:ordfield:ii] For \(x,y \in F\), \(x > 0\) and \(y > 0\) implies \(xy > 0\).

If \(x > 0\), we say \(x\) is *positive*. If \(x < 0\), we say \(x\) is *negative*. We also say \(x\) is *nonnegative* if \(x \geq 0\), and \(x\) is *nonpositive* if \(x \leq 0\).

For example, it can be checked that the rational numbers \({\mathbb{Q}}\) with the standard ordering is an ordered field.

[prop:bordfield] Let \(F\) be an ordered field and \(x,y,z \in F\). Then:

- [prop:bordfield:i] If \(x > 0\), then \(-x < 0\) (and vice-versa).
- [prop:bordfield:ii] If \(x > 0\) and \(y < z\), then \(xy < xz\).
- [prop:bordfield:iii] If \(x < 0\) and \(y < z\), then \(xy > xz\).
- [prop:bordfield:iv] If \(x \not= 0\), then \(x^2 > 0\).
- [prop:bordfield:v] If \(0 < x < y\), then \(0 < \nicefrac{1}{y} < \nicefrac{1}{x}\).

Note that [prop:bordfield:iv] implies in particular that \(1 > 0\).

Let us prove [prop:bordfield:i]. The inequality \(x > 0\) implies by item [defn:ordfield:i] of definition of ordered field that \(x + (-x) > 0 + (-x)\). Now apply the algebraic properties of fields to obtain \(0 > -x\). The “vice-versa” follows by similar calculation.

For [prop:bordfield:ii], first notice that \(y < z\) implies \(0 < z - y\) by applying item [defn:ordfield:i] of the definition of ordered fields. Now apply item [defn:ordfield:ii] of the definition of ordered fields to obtain \(0 < x(z-y)\). By algebraic properties we get \(0 < xz - xy\), and again applying item [defn:ordfield:i] of the definition we obtain \(xy < xz\).

Part [prop:bordfield:iii] is left as an exercise.

To prove part [prop:bordfield:iv] first suppose \(x > 0\). Then by item [defn:ordfield:ii] of the definition of ordered fields we obtain that \(x^2 > 0\) (use \(y=x\)). If \(x < 0\), we use part [prop:bordfield:iii] of this proposition. Plug in \(y=x\) and \(z=0\).

Finally to prove part [prop:bordfield:v], notice that \(\nicefrac{1}{x}\) cannot be equal to zero (why?). Suppose \(\nicefrac{1}{x} < 0\), then \(\nicefrac{-1}{x} > 0\) by [prop:bordfield:i]. Then apply part [prop:bordfield:ii] (as \(x > 0\)) to obtain \(x(\nicefrac{-1}{x}) > 0x\) or \(-1 > 0\), which contradicts \(1 > 0\) by using part [prop:bordfield:i] again. Hence \(\nicefrac{1}{x} > 0\). Similarly \(\nicefrac{1}{y} > 0\). Thus \((\nicefrac{1}{x})(\nicefrac{1}{y}) > 0\) by definition of ordered field and by part [prop:bordfield:ii] \[(\nicefrac{1}{x})(\nicefrac{1}{y})x < (\nicefrac{1}{x})(\nicefrac{1}{y})y .\] By algebraic properties we get \(\nicefrac{1}{y} < \nicefrac{1}{x}\).

Product of two positive numbers (elements of an ordered field) is positive. However, it is not true that if the product is positive, then each of the two factors must be positive.

Let \(x,y \in F\) where \(F\) is an ordered field. Suppose \(xy > 0\). Then either both \(x\) and \(y\) are positive, or both are negative.

Clearly both of the conclusions can happen. If either \(x\) and \(y\) are zero, then \(xy\) is zero and hence not positive. Hence we assume that \(x\) and \(y\) are nonzero, and we simply need to show that if they have opposite signs, then \(xy < 0\). Without loss of generality suppose \(x > 0\) and \(y < 0\). Multiply \(y < 0\) by \(x\) to get \(xy < 0x = 0\). The result follows by contrapositive.

The reader may also know about the *complex numbers*, usually denoted by \({\mathbb{C}}\). That is, \({\mathbb{C}}\) is the set of numbers of the form \(x + iy\), where \(x\) and \(y\) are real numbers, and \(i\) is the imaginary number, a number such that \(i^2 = -1\). The reader may remember from algebra that \({\mathbb{C}}\) is also a field, however, it is not an ordered field. While one can make \({\mathbb{C}}\) into an ordered set in some way, it can be proved that it is not possible to put an order on \({\mathbb{C}}\) that will make it an ordered field.

## Exercises

Prove part [prop:bordfield:iii] of .

[exercise:finitesethasminmax] Let \(S\) be an ordered set. Let \(A \subset S\) be a nonempty finite subset. Then \(A\) is bounded. Furthermore, \(\inf\, A\) exists and is in \(A\) and \(\sup\, A\) exists and is in \(A\). Hint: Use .

Let \(x, y \in F\), where \(F\) is an ordered field. Suppose \(0 < x < y\). Show that \(x^2 < y^2\).

Let \(S\) be an ordered set. Let \(B \subset S\) be bounded (above and below). Let \(A \subset B\) be a nonempty subset. Suppose all the \(\inf\)’s and \(\sup\)’s exist. Show that \[\inf\, B \leq \inf\, A \leq \sup\, A \leq \sup\, B .\]

Let \(S\) be an ordered set. Let \(A \subset S\) and suppose \(b\) is an upper bound for \(A\). Suppose \(b \in A\). Show that \(b = \sup\, A\).

Let \(S\) be an ordered set. Let \(A \subset S\) be a nonempty subset that is bounded above. Suppose \(\sup\, A\) exists and \(\sup\, A \notin A\). Show that \(A\) contains a countably infinite subset. In particular, \(A\) is infinite.

Find a (nonstandard) ordering of the set of natural numbers \({\mathbb{N}}\) such that there exists a nonempty proper subset \(A \subsetneq {\mathbb{N}}\) and such that \(\sup\, A\) exists in \({\mathbb{N}}\), but \(\sup\, A \notin A\).

Let \(F = \{ 0, 1, 2 \}\). a) Prove that there is exactly one way to define addition and multiplication so that \(F\) is a field if \(0\) and \(1\) have their usual meaning of (A4) and (M4). b) Show that \(F\) cannot be an ordered field.

[exercise:dominatingb] Let \(S\) be an ordered set and \(A\) is a nonempty subset such that \(\sup \, A\) exists. Suppose there is a \(B \subset A\) such that whenever \(x \in A\) there is a \(y \in B\) such that \(x \leq y\). Show that \(\sup \, B\) exists and \(\sup \, B = \sup \, A\).

Let \(D\) be the ordered set of all possible words (not just English words, all strings of letters of arbitrary length) using the Latin alphabet using only lower case letters. The order is the lexicographic order as in a dictionary (e.g. *aaa* \(<\) *dog* \(<\) *door*). Let \(A\) be the subset of \(D\) containing the words whose first letter is ‘a’ (e.g. *a* \(\in A\), *abcd* \(\in A\)). Show that \(A\) has a supremum and find what it is.

### The set of real numbers

*Note: 2 lectures, the extended real numbers are optional*

#### The set of real numbers

We finally get to the real number system. To simplify matters, instead of constructing the real number set from the rational numbers, we simply state their existence as a theorem without proof. Notice that \({\mathbb{Q}}\) is an ordered field.

There exists a unique^{4} ordered field \({\mathbb{R}}\) with the such that \({\mathbb{Q}}\subset {\mathbb{R}}\).

Note that also \({\mathbb{N}}\subset {\mathbb{Q}}\). We saw that \(1 > 0\). By (exercise) we can prove that \(n > 0\) for all \(n \in {\mathbb{N}}\). Similarly we can easily verify all the statements we know about rational numbers and their natural ordering.

Let us prove one of the most basic but useful results about the real numbers. The following proposition is essentially how an analyst proves that a number is zero.

If \(x \in {\mathbb{R}}\) is such that \(x \geq 0\) and \(x \leq \epsilon\) for all \(\epsilon \in {\mathbb{R}}\) where \(\epsilon > 0\), then \(x = 0\).

If \(x > 0\), then \(0 < \nicefrac{x}{2} < x\) (why?). Taking \(\epsilon = \nicefrac{x}{2}\) obtains a contradiction. Thus \(x=0\).

A related simple fact is that any time we have two real numbers \(a < b\), then there is another real number \(c\) such that \(a < c < b\). Just take for example \(c = \frac{a+b}{2}\) (why?). In fact, there are infinitely many real numbers between \(a\) and \(b\).

The most useful property of \({\mathbb{R}}\) for analysts is not just that it is an ordered field, but that it has the . Essentially we want \({\mathbb{Q}}\), but we also want to take suprema (and infima) willy-nilly. So what we do is to throw in enough numbers to obtain \({\mathbb{R}}\).

We mentioned already that \({\mathbb{R}}\) must contain elements that are not in \({\mathbb{Q}}\) because of the . We saw there is no rational square root of two. The set \(\{ x \in {\mathbb{Q}}: x^2 < 2 \}\) implies the existence of the real number \(\sqrt{2}\), although this fact requires a bit of work.

[example:sqrt2] Claim: There exists a unique positive real number \(r\) such that \(r^2 = 2\). We denote \(r\) by \(\sqrt{2}\).

Take the set \(A := \{ x \in {\mathbb{R}}: x^2 < 2 \}\). First if \(x^2 < 2\), then \(x < 2\). To see this fact, note that \(x \geq 2\) implies \(x^2 \geq 4\) (use , we will not explicitly mention its use from now on), hence any number \(x\) such that \(x \geq 2\) is not in \(A\). Thus \(A\) is bounded above. On the other hand, \(1 \in A\), so \(A\) is nonempty.

Let us define \(r := \sup\, A\). We will show that \(r^2 = 2\) by showing that \(r^2 \geq 2\) and \(r^2 \leq 2\). This is the way analysts show equality, by showing two inequalities. We already know that \(r \geq 1 > 0\).

In the following, it may seem we are pulling certain expressions out of a hat. When writing a proof such as this we would, of course, come up with the expressions only after playing around with what we wish to prove. The order in which we write the proof is not necessarily the order in which we come up with the proof.

Let us first show that \(r^2 \geq 2\). Take a positive number \(s\) such that \(s^2 < 2\). We wish to find an \(h > 0\) such that \({(s+h)}^2 < 2\). As \(2-s^2 > 0\), we have \(\frac{2-s^2}{2s+1} > 0\). We choose an \(h \in {\mathbb{R}}\) such that \(0 < h < \frac{2-s^2}{2s+1}\). Furthermore, we assume \(h < 1\).

\[\begin{aligned} {(s+h)}^2 - s^2 & = h(2s + h) \\ & < h(2s+1) & & \quad \,\bigl(\text{since } h < 1\bigr) \\ & < 2-s^2 & & \quad \left(\text{since } h < \frac{2-s^2}{2s+1} \right) . \end{aligned}\]

Therefore, \({(s+h)}^2 < 2\). Hence \(s+h \in A\), but as \(h > 0\) we have \(s+h > s\). So \(s < r = \sup\, A\). As \(s\) was an arbitrary positive number such that \(s^2 < 2\), it follows that \(r^2 \geq 2\).

Now take a positive number \(s\) such that \(s^2 > 2\). We wish to find an \(h > 0\) such that \({(s-h)}^2 > 2\). As \(s^2-2 > 0\) we have \(\frac{s^2-2}{2s} > 0\). We choose an \(h \in {\mathbb{R}}\) such that \(0 < h < \frac{s^2-2}{2s}\) and \(h < s\). \[\begin{split} s^2 - {(s-h)}^2 & = 2sh - h^2 \\ & < 2sh \\ & < s^2-2 \qquad \left( \text{since } h < \frac{s^2-2}{2s} \right) . \end{split}\] By subtracting \(s^2\) from both sides and multiplying by \(-1\), we find \({(s-h)}^2 > 2\). Therefore \(s-h \notin A\).

Furthermore, if \(x \geq s-h\), then \(x^2 \geq {(s-h)}^2 > 2\) (as \(x > 0\) and \(s-h > 0\)) and so \(x \notin A\). Thus \(s-h\) is an upper bound for \(A\). However, \(s-h < s\), or in other words \(s > r = \sup\, A\). Thus \(r^2 \leq 2\).

Together, \(r^2 \geq 2\) and \(r^2 \leq 2\) imply \(r^2 = 2\). The existence part is finished. We still need to handle uniqueness. Suppose \(s \in {\mathbb{R}}\) such that \(s^2 = 2\) and \(s > 0\). Thus \(s^2 = r^2\). However, if \(0 < s < r\), then \(s^2 < r^2\). Similarly \(0 < r < s\) implies \(r^2 < s^2\). Hence \(s = r\).

The number \(\sqrt{2} \notin {\mathbb{Q}}\). The set \({\mathbb{R}}\setminus {\mathbb{Q}}\) is called the set of *irrational* numbers. We just saw that \({\mathbb{R}}\setminus {\mathbb{Q}}\) is nonempty. Not only is it nonempty, we will see later that is it very large indeed.

Using the same technique as above, we can show that a positive real number \(x^{1/n}\) exists for all \(n\in {\mathbb{N}}\) and all \(x > 0\). That is, for each \(x > 0\), there exists a unique positive real number \(r\) such that \(r^n = x\). The proof is left as an exercise.

#### Archimedean property

As we have seen, there are plenty of real numbers in any interval. But there are also infinitely many rational numbers in any interval. The following is one of the fundamental facts about the real numbers. The two parts of the next theorem are actually equivalent, even though it may not seem like that at first sight.

[thm:arch]

- [thm:arch:i]
*(Archimedean property)*If \(x, y \in {\mathbb{R}}\) and \(x > 0\), then there exists an \(n \in {\mathbb{N}}\) such that \[nx > y .\] - [thm:arch:ii]
*(\({\mathbb{Q}}\) is dense in \({\mathbb{R}}\))*If \(x, y \in {\mathbb{R}}\) and \(x < y\), then there exists an \(r \in {\mathbb{Q}}\) such that \(x < r < y\).

Let us prove [thm:arch:i]. We divide through by \(x\) and then [thm:arch:i] says that for any real number \(t:= \nicefrac{y}{x}\), we can find natural number \(n\) such that \(n > t\). In other words, [thm:arch:i] says that \({\mathbb{N}}\subset {\mathbb{R}}\) is not bounded above. Suppose for contradiction that \({\mathbb{N}}\) is bounded above. Let \(b := \sup {\mathbb{N}}\). The number \(b-1\) cannot possibly be an upper bound for \({\mathbb{N}}\) as it is strictly less than \(b\) (the supremum). Thus there exists an \(m \in {\mathbb{N}}\) such that \(m > b-1\). We add one to obtain \(m+1 > b\), which contradicts \(b\) being an upper bound.

Let us tackle [thm:arch:ii]. First assume \(x \geq 0\). Note that \(y-x > 0\). By [thm:arch:i], there exists an \(n \in {\mathbb{N}}\) such that \[n(y-x) > 1.\] Also by [thm:arch:i] the set \(A := \{ k \in {\mathbb{N}}: k > nx \}\) is nonempty. By the well ordering property of \({\mathbb{N}}\), \(A\) has a least element \(m\). As \(m \in A\), then \(m > nx\). We divide through by \(n\) to get \(x < \nicefrac{m}{n}\). As \(m\) is the least element of \(A\), \(m-1 \notin A\). If \(m > 1\), then \(m-1 \in {\mathbb{N}}\), but \(m-1 \notin A\) and so \(m-1 \leq nx\). If \(m=1\), then \(m-1 = 0\), and \(m-1 \leq nx\) still holds as \(x \geq 0\). In other words, \[m-1 \leq nx \qquad \text{or} \qquad m \leq nx+1 . % < m .\] On the other hand from \(n(y-x) > 1\) we obtain \(ny > 1+nx\). Hence \(ny > 1+nx \geq m\), and therefore \(y > \nicefrac{m}{n}\). Putting everything together we obtain \(x < \nicefrac{m}{n} < y\). So let \(r = \nicefrac{m}{n}\).

Now assume \(x < 0\). If \(y > 0\), then we just take \(r=0\). If \(y \leq 0\), then \(0 \leq -y < -x\), and we find a rational \(q\) such that \(-y < q < -x\). Then take \(r = -q\).

Let us state and prove a simple but useful corollary of the .

\(\inf \{ \nicefrac{1}{n} : n \in {\mathbb{N}}\} = 0\).

Let \(A := \{ \nicefrac{1}{n} : n \in {\mathbb{N}}\}\). Obviously \(A\) is not empty. Furthermore, \(\nicefrac{1}{n} > 0\) and so 0 is a lower bound, and \(b := \inf\, A\) exists. As 0 is a lower bound, then \(b \geq 0\). Now take an arbitrary \(a > 0\). By the there exists an \(n\) such that \(na > 1\), or in other words \(a > \nicefrac{1}{n} \in A\). Therefore \(a\) cannot be a lower bound for \(A\). Hence \(b=0\).

#### Using supremum and infimum

We want to make sure that suprema and infima are compatible with algebraic operations. For a set \(A \subset {\mathbb{R}}\) and a number \(x \in {\mathbb{R}}\) define \[\begin{aligned} x + A & := \{ x+y \in {\mathbb{R}}: y \in A \} , \\ xA & := \{ xy \in {\mathbb{R}}: y \in A \} .\end{aligned}\]

[prop:supinfalg] Let \(A \subset {\mathbb{R}}\) be nonempty.

- If \(x \in {\mathbb{R}}\) and \(A\) is bounded above, then \(\sup (x+A) = x + \sup\, A\).
- If \(x \in {\mathbb{R}}\) and \(A\) is bounded below, then \(\inf (x+A) = x + \inf\, A\).
- If \(x > 0\) and \(A\) is bounded above, then \(\sup (xA) = x ( \sup\, A )\).
- If \(x > 0\) and \(A\) is bounded below, then \(\inf (xA) = x ( \inf\, A )\).
- If \(x < 0\) and \(A\) is bounded below, then \(\sup (xA) = x ( \inf\, A )\).
- If \(x < 0\) and \(A\) is bounded above, then \(\inf (xA) = x ( \sup\, A )\).

Do note that multiplying a set by a negative number switches supremum for an infimum and vice-versa. Also, as the proposition implies that supremum (resp. infimum) of \(x+A\) or \(xA\) exists, it also implies that \(x+A\) or \(xA\) is nonempty and bounded from above (resp. from below).

Let us only prove the first statement. The rest are left as exercises.

Suppose \(b\) is an upper bound for \(A\). That is, \(y \leq b\) for all \(y \in A\). Then \(x+y \leq x+b\) for all \(y \in A\), and so \(x+b\) is an upper bound for \(x+A\). In particular, if \(b = \sup\, A\), then \[\sup (x+A) \leq x+b = x+ \sup\, A .\]

The other direction is similar. If \(b\) is an upper bound for \(x+A\), then \(x+y \leq b\) for all \(y \in A\) and so \(y \leq b-x\) for all \(y \in A\). So \(b-x\) is an upper bound for \(A\). If \(b = \sup (x+A)\), then \[\sup\, A \leq b-x = \sup (x+A) -x .\] And the result follows.

Sometimes we need to apply supremum or infimum twice. Here is an example.

[infsupineq:prop] Let \(A, B \subset {\mathbb{R}}\) be nonempty sets such that \(x \leq y\) whenever \(x \in A\) and \(y \in B\). Then \(A\) is bounded above, \(B\) is bounded below, and \(\sup\, A \leq \inf\, B\).

Any \(x \in A\) is a lower bound for \(B\). Therefore \(x \leq \inf\, B\) for all \(x \in A\), so \(\inf\, B\) is an upper bound for \(A\). Hence, \(\sup\, A \leq \inf\, B\).

We must be careful about strict inequalities and taking suprema and infima. Note that \(x < y\) whenever \(x \in A\) and \(y \in B\) still only implies \(\sup\, A \leq \inf\, B\), and not a strict inequality. This is an important subtle point that comes up often. For example, take \(A := \{ 0 \}\) and take \(B := \{ \nicefrac{1}{n} : n \in {\mathbb{N}}\}\). Then \(0 < \nicefrac{1}{n}\) for all \(n \in {\mathbb{N}}\). However, \(\sup\, A = 0\) and \(\inf\, B = 0\).

The proof of the following often used elementary fact is left to the reader. A similar statement holds for infima.

[prop:existsxepsfromsup] If \(S \subset {\mathbb{R}}\) is a nonempty set, bounded from above, then for every \(\epsilon > 0\) there exists \(x \in S\) such that \((\sup\, S) - \epsilon < x \leq \sup\, S\).

To make using suprema and infima even easier, we may want to write \(\sup\, A\) and \(\inf\, A\) without worrying about \(A\) being bounded and nonempty. We make the following natural definitions.

Let \(A \subset {\mathbb{R}}\) be a set.

- If \(A\) is empty, then \(\sup\, A := -\infty\).
- If \(A\) is not bounded above, then \(\sup\, A := \infty\).
- If \(A\) is empty, then \(\inf\, A := \infty\).
- If \(A\) is not bounded below, then \(\inf\, A := -\infty\).

For convenience, \(\infty\) and \(-\infty\) are sometimes treated as if they were numbers, except we do not allow arbitrary arithmetic with them. We make \({\mathbb{R}}^* := {\mathbb{R}}\cup \{ -\infty , \infty\}\) into an ordered set by letting \[-\infty < \infty \quad \text{and} \quad -\infty < x \quad \text{and} \quad x < \infty \quad \text{for all $x \in {\mathbb{R}}$}.\] The set \({\mathbb{R}}^*\) is called the set of *extended real numbers*. It is possible to define some arithmetic on \({\mathbb{R}}^*\). Most operations are extended in an obvious way, but we must leave \(\infty-\infty\), \(0 \cdot (\pm\infty)\), and \(\frac{\pm\infty}{\pm\infty}\) undefined. We refrain from using this arithmetic, it leads to easy mistakes as \({\mathbb{R}}^*\) is not a field. Now we can take suprema and infima without fear of emptiness or unboundedness. In this book we mostly avoid using \({\mathbb{R}}^*\) outside of exercises, and leave such generalizations to the interested reader.

#### Maxima and minima

By we know a finite set of numbers always has a supremum or an infimum that is contained in the set itself. In this case we usually do not use the words supremum or infimum.

When a set \(A\) of real numbers is bounded above, such that \(\sup\, A \in A\), then we can use the word *maximum* and the notation \(\max A\) to denote the supremum. Similarly for infimum; when a set \(A\) is bounded below and \(\inf\, A \in A\), then we can use the word *minimum* and the notation \(\min\, A\). For example, \[\begin{aligned} & \max \{ 1,2.4,\pi,100 \} = 100 , \\ & \min \{ 1,2.4,\pi,100 \} = 1 .\end{aligned}\] While writing \(\sup\) and \(\inf\) may be technically correct in this situation, \(\max\) and \(\min\) are generally used to emphasize that the supremum or infimum is in the set itself.

## Exercises

Prove that if \(t > 0\) (\(t \in {\mathbb{R}}\)), then there exists an \(n \in {\mathbb{N}}\) such that \(\dfrac{1}{n^2} < t\).

Prove that if \(t \geq 0\) (\(t \in {\mathbb{R}}\)), then there exists an \(n \in {\mathbb{N}}\) such that \(n-1 \leq t < n\).

Finish the proof of .

Let \(x, y \in {\mathbb{R}}\). Suppose \(x^2 + y^2 = 0\). Prove that \(x = 0\) and \(y = 0\).

Show that \(\sqrt{3}\) is irrational.

Let \(n \in {\mathbb{N}}\). Show that either \(\sqrt{n}\) is either an integer or it is irrational.

Prove the *arithmetic-geometric mean inequality*. That is, for two positive real numbers \(x,y\) we have \[\sqrt{xy} \leq \frac{x+y}{2} .\] Furthermore, equality occurs if and only if \(x=y\).

Show that for any two real numbers \(x\) and \(y\) such that \(x < y\), there exists an irrational number \(s\) such that \(x < s < y\). Hint: Apply the density of \({\mathbb{Q}}\) to \(\dfrac{x}{\sqrt{2}}\) and \(\dfrac{y}{\sqrt{2}}\).

[exercise:supofsum] Let \(A\) and \(B\) be two nonempty bounded sets of real numbers. Let \(C := \{ a+b : a \in A, b \in B \}\). Show that \(C\) is a bounded set and that \[\sup\,C = \sup\,A + \sup\,B \qquad \text{and} \qquad \inf\,C = \inf\,A + \inf\,B .\]

Let \(A\) and \(B\) be two nonempty bounded sets of nonnegative real numbers. Define the set \(C := \{ ab : a \in A, b \in B \}\). Show that \(C\) is a bounded set and that \[\sup\,C = (\sup\,A )( \sup\,B) \qquad \text{and} \qquad \inf\,C = (\inf\,A )( \inf\,B).\]

Given \(x > 0\) and \(n \in {\mathbb{N}}\), show that there exists a unique positive real number \(r\) such that \(x = r^n\). Usually \(r\) is denoted by \(x^{1/n}\).

Prove .

[exercise:bernoulliineq] Prove the so-called *Bernoulli’s inequality* ^{5} : If \(1+x > 0\) then for all \(n \in {\mathbb{N}}\) we have \((1+x)^n \geq 1+nx\).

### Absolute value

*Note: 0.5–1 lecture*

A concept we will encounter over and over is the concept of *absolute value*. You want to think of the absolute value as the “size” of a real number. Let us give a formal definition. \[\left\lvert {x} \right\rvert := \begin{cases} x & \text{ if $x \geq 0$}, \\ -x & \text{ if $x < 0$} . \end{cases}\]

Let us give the main features of the absolute value as a proposition.

[prop:absbas]

- [prop:absbas:i] \(\left\lvert {x} \right\rvert \geq 0\), and \(\left\lvert {x} \right\rvert=0\) if and only if \(x = 0\).
- [prop:absbas:ii] \(\left\lvert {-x} \right\rvert = \left\lvert {x} \right\rvert\) for all \(x \in {\mathbb{R}}\).
- [prop:absbas:iii] \(\left\lvert {xy} \right\rvert = \left\lvert {x} \right\rvert\left\lvert {y} \right\rvert\) for all \(x,y \in {\mathbb{R}}\).
- [prop:absbas:iv] \(\left\lvert {x} \right\rvert^2 = x^2\) for all \(x \in {\mathbb{R}}\).
- [prop:absbas:v] \(\left\lvert {x} \right\rvert \leq y\) if and only if \(-y \leq x \leq y\).
- [prop:absbas:vi] \(-\left\lvert {x} \right\rvert \leq x \leq \left\lvert {x} \right\rvert\) for all \(x \in {\mathbb{R}}\).

[prop:absbas:i]: This statement is not difficult to see from the definition.

[prop:absbas:ii]: Suppose \(x > 0\), then \(\left\lvert {-x} \right\rvert = -(-x) = x = \left\lvert {x} \right\rvert\). Similarly when \(x < 0\), or \(x = 0\).

[prop:absbas:iii]: If \(x\) or \(y\) is zero, then the result is obvious. When \(x\) and \(y\) are both positive, then \(\left\lvert {x} \right\rvert\left\lvert {y} \right\rvert = xy\). \(xy\) is also positive and hence \(xy = \left\lvert {xy} \right\rvert\). If \(x\) and \(y\) are both negative then \(xy\) is still positive and \(xy = \left\lvert {xy} \right\rvert\), and \(\left\lvert {x} \right\rvert\left\lvert {y} \right\rvert = (-x)(-y) = xy\). Next assume \(x > 0\) and \(y < 0\). Then \(\left\lvert {x} \right\rvert\left\lvert {y} \right\rvert = x(-y) = -(xy)\). Now \(xy\) is negative and hence \(\left\lvert {xy} \right\rvert = -(xy)\). Similarly if \(x < 0\) and \(y > 0\).

[prop:absbas:iv]: Obvious if \(x \geq 0\). If \(x < 0\), then \(\left\lvert {x} \right\rvert^2 = {(-x)}^2 = x^2\).

[prop:absbas:v]: Suppose \(\left\lvert {x} \right\rvert \leq y\). If \(x \geq 0\), then \(x \leq y\). Obviously \(y \geq 0\) and hence \(-y \leq 0 \leq x\) so \(-y \leq x \leq y\) holds. If \(x < 0\), then \(\left\lvert {x} \right\rvert \leq y\) means \(-x \leq y\). Negating both sides we get \(x \geq -y\). Again \(y \geq 0\) and so \(y \geq 0 > x\). Hence, \(-y \leq x \leq y\).

On the other hand, suppose \(-y \leq x \leq y\) is true. If \(x \geq 0\), then \(x \leq y\) is equivalent to \(\left\lvert {x} \right\rvert \leq y\). If \(x < 0\), then \(-y \leq x\) implies \((-x) \leq y\), which is equivalent to \(\left\lvert {x} \right\rvert \leq y\).

[prop:absbas:vi]: Apply [prop:absbas:v] with \(y = \left\lvert {x} \right\rvert\).

A property used frequently enough to give it a name is the so-called *triangle inequality*.

\(\left\lvert {x+y} \right\rvert \leq \left\lvert {x} \right\rvert+\left\lvert {y} \right\rvert\) for all \(x, y \in {\mathbb{R}}\).

From we have \(- \left\lvert {x} \right\rvert \leq x \leq \left\lvert {x} \right\rvert\) and \(- \left\lvert {y} \right\rvert \leq y \leq \left\lvert {y} \right\rvert\). We add these two inequalities to obtain \[- (\left\lvert {x} \right\rvert+\left\lvert {y} \right\rvert) \leq x+y \leq \left\lvert {x} \right\rvert+ \left\lvert {y} \right\rvert .\] Again by we have \(\left\lvert {x+y} \right\rvert \leq \left\lvert {x} \right\rvert+\left\lvert {y} \right\rvert\).

There are other often applied versions of the triangle inequality.

Let \(x,y \in {\mathbb{R}}\)

*(reverse triangle inequality)*\(\bigl\lvert (\left\lvert {x} \right\rvert-\left\lvert {y} \right\rvert) \bigr\rvert \leq \left\lvert {x-y} \right\rvert\).- \(\left\lvert {x-y} \right\rvert \leq \left\lvert {x} \right\rvert+\left\lvert {y} \right\rvert\).

Let us plug in \(x=a-b\) and \(y=b\) into the standard triangle inequality to obtain \[\left\lvert {a} \right\rvert = \left\lvert {a-b+b} \right\rvert \leq \left\lvert {a-b} \right\rvert + \left\lvert {b} \right\rvert ,\] or \(\left\lvert {a} \right\rvert-\left\lvert {b} \right\rvert \leq \left\lvert {a-b} \right\rvert\). Switching the roles of \(a\) and \(b\) we obtain or \(\left\lvert {b} \right\rvert-\left\lvert {a} \right\rvert \leq \left\lvert {b-a} \right\rvert = \left\lvert {a-b} \right\rvert\). Now applying again we obtain the reverse triangle inequality.

The second version of the triangle inequality is obtained from the standard one by just replacing \(y\) with \(-y\) and noting again that \(\left\lvert {-y} \right\rvert = \left\lvert {y} \right\rvert\).

Let \(x_1, x_2, \ldots, x_n \in {\mathbb{R}}\). Then \[\left\lvert {x_1 + x_2 + \cdots + x_n} \right\rvert \leq \left\lvert {x_1} \right\rvert + \left\lvert {x_2} \right\rvert + \cdots + \left\lvert {x_n} \right\rvert .\]

We proceed by . The conclusion holds trivially for \(n=1\), and for \(n=2\) it is the standard triangle inequality. Suppose the corollary holds for \(n\). Take \(n+1\) numbers \(x_1,x_2,\ldots,x_{n+1}\) and first use the standard triangle inequality, then the induction hypothesis \[\begin{split} \lvert {x_1 + x_2 + \cdots + x_n + x_{n+1}} \rvert & \leq \lvert {x_1 + x_2 + \cdots + x_n} \rvert + \lvert {x_{n+1}} \rvert \\ & \leq \lvert {x_1} \rvert + \lvert {x_2} \rvert + \cdots + \lvert {x_n} \rvert + \lvert {x_{n+1}} \rvert . \qedhere \end{split}\]

Let us see an example of the use of the triangle inequality.

Find a number \(M\) such that \(\lvert {x^2-9x+1} \rvert \leq M\) for all \(-1 \leq x \leq 5\).

Using the triangle inequality, write \[\lvert {x^2-9x+1} \rvert \leq \lvert {x^2} \rvert+\lvert {9x} \rvert+\lvert {1} \rvert = \lvert {x} \rvert^2+9\lvert {x} \rvert+1 .\] It is obvious that \(\lvert {x} \rvert^2+9\lvert {x} \rvert+1\) is largest when \(\left\lvert {x} \right\rvert\) is largest. In the interval provided, \(\left\lvert {x} \right\rvert\) is largest when \(x=5\) and so \(\left\lvert {x} \right\rvert=5\). One possibility for \(M\) is \[M = 5^2+9(5)+1 = 71 .\] There are, of course, other \(M\) that work. The bound of 71 is much higher than it need be, but we didn’t ask for the best possible \(M\), just one that works.

The last example leads us to the concept of bounded functions.

Suppose \(f \colon D \to {\mathbb{R}}\) is a function. We say \(f\) is *bounded* if there exists a number \(M\) such that \(\left\lvert {f(x)} \right\rvert \leq M\) for all \(x \in D\).

In the example we proved \(x^2-9x+1\) is bounded when considered as a function on \(D = \{ x : -1 \leq x \leq 5 \}\). On the other hand, if we consider the same polynomial as a function on the whole real line \({\mathbb{R}}\), then it is not bounded.

For a function \(f \colon D \to {\mathbb{R}}\) we write \[\begin{aligned} & \sup_{x \in D} f(x) := \sup\, f(D) , \\ & \inf_{x \in D} f(x) := \inf\, f(D) .\end{aligned}\] We also sometimes replace the “\(x \in D\)” with an expression. For example if, as before, \(f(x) = x^2-9x+1\), for \(-1 \leq x \leq 5\), a little bit of calculus shows \[\sup_{x \in D} f(x) = \sup_{-1 \leq x \leq 5} ( x^2 -9x+1 ) = 11, \qquad \inf_{x \in D} f(x) = \inf_{-1 \leq x \leq 5} ( x^2 -9x+1 ) = \nicefrac{-77}{4} .\]

[prop:funcsupinf] If \(f \colon D \to {\mathbb{R}}\) and \(g \colon D \to {\mathbb{R}}\) (\(D\) nonempty) are bounded^{6} functions and \[f(x) \leq g(x) \qquad \text{for all $x \in D$},\] then \[\label{prop:funcsupinf:eq} \sup_{x \in D} f(x) \leq \sup_{x \in D} g(x) \qquad \text{and} \qquad \inf_{x \in D} f(x) \leq \inf_{x \in D} g(x) .\]

You should be careful with the variables. The \(x\) on the left side of the inequality in [prop:funcsupinf:eq] is different from the \(x\) on the right. You should really think of the first inequality as \[\sup_{x \in D} f(x) \leq \sup_{y \in D} g(y) .\] Let us prove this inequality. If \(b\) is an upper bound for \(g(D)\), then \(f(x) \leq g(x) \leq b\) for all \(x \in D\), and hence \(b\) is an upper bound for \(f(D)\). Taking the least upper bound we get that for all \(x \in D\) \[f(x) \leq \sup_{y \in D} g(y) .\] Therefore \(\sup_{y \in D} g(y)\) is an upper bound for \(f(D)\) and thus greater than or equal to the least upper bound of \(f(D)\). \[\sup_{x \in D} f(x) \leq \sup_{y \in D} g(y) .\] The second inequality (the statement about the inf) is left as an exercise.

A common mistake is to conclude \[\label{rn:av:ltnottrue} \sup_{x \in D} f(x) \leq \inf_{y \in D} g(y) .\] The inequality [rn:av:ltnottrue] is not true given the hypothesis of the claim above. For this stronger inequality we need the stronger hypothesis \[f(x) \leq g(y) \qquad \text{for all $x \in D$ and $y \in D$.}\] The proof as well as a counterexample is left as an exercise.

## Exercises

Show that \(\left\lvert {x-y} \right\rvert < \epsilon\) if and only if \(x-\epsilon < y < x+\epsilon\).

Show that

- \(\max \{x,y\} = \frac{x+y+\left\lvert {x-y} \right\rvert}{2}\)
- \(\min \{x,y\} = \frac{x+y-\left\lvert {x-y} \right\rvert}{2}\)

Find a number \(M\) such that \(\lvert {x^3-x^2+8x} \rvert \leq M\) for all \(-2 \leq x \leq 10\).

Finish the proof of . That is, prove that given any set \(D\), and two bounded functions \(f \colon D \to {\mathbb{R}}\) and \(g \colon D \to {\mathbb{R}}\) such that \(f(x) \leq g(x)\) for all \(x \in D\), then \[\inf_{x\in D} f(x) \leq \inf_{x\in D} g(x) .\]

Let \(f \colon D \to {\mathbb{R}}\) and \(g \colon D \to {\mathbb{R}}\) be functions (\(D\) nonempty).

- Suppose \(f(x) \leq g(y)\) for all \(x \in D\) and \(y \in D\). Show that \[\sup_{x\in D} f(x) \leq \inf_{x\in D} g(x) .\]
- Find a specific \(D\), \(f\), and \(g\), such that \(f(x) \leq g(x)\) for all \(x \in D\), but \[\sup_{x\in D} f(x) > \inf_{x\in D} g(x) .\]

Prove without the assumption that the functions are bounded. Hint: You need to use the extended real numbers.

[exercise:sumofsup] Let \(D\) be a nonempty set. Suppose \(f \colon D \to {\mathbb{R}}\) and \(g \colon D \to {\mathbb{R}}\) are bounded functions. a) Show \[\sup_{x\in D} \bigl(f(x) + g(x) \bigr) \leq \sup_{x\in D} f(x) + \sup_{x\in D} g(x) \qquad \text{and} \qquad \inf_{x\in D} \bigl(f(x) + g(x) \bigr) \geq \inf_{x\in D} f(x) + \inf_{x\in D} g(x) .\] b) Find examples where we obtain strict inequalities.

### Intervals and the size of R

*Note: 0.5–1 lecture (proof of uncountability of \({\mathbb{R}}\) can be optional)*

You surely saw the notation for intervals before, but let us give a formal definition here. For \(a, b \in {\mathbb{R}}\) such that \(a < b\) we define \[\begin{aligned} & [a,b] := \{ x \in {\mathbb{R}}: a \leq x \leq b \}, \\ & (a,b) := \{ x \in {\mathbb{R}}: a < x < b \}, \\ & (a,b] := \{ x \in {\mathbb{R}}: a < x \leq b \}, \\ & [a,b) := \{ x \in {\mathbb{R}}: a \leq x < b \} .\end{aligned}\] The interval \([a,b]\) is called a *closed interval* and \((a,b)\) is called an *open interval*. The intervals of the form \((a,b]\) and \([a,b)\) are called *half-open intervals*.

The above intervals were all *bounded intervals*, since both \(a\) and \(b\) were real numbers. We define *unbounded intervals*, \[\begin{aligned} & [a,\infty) := \{ x \in {\mathbb{R}}: a \leq x \}, \\ & (a,\infty) := \{ x \in {\mathbb{R}}: a < x \}, \\ & (-\infty,b] := \{ x \in {\mathbb{R}}: x \leq b \}, \\ & (-\infty,b) := \{ x \in {\mathbb{R}}: x < b \} .\end{aligned}\] For completeness we define \((-\infty,\infty) := {\mathbb{R}}\).

In short, an interval is a set \(I \subset {\mathbb{R}}\) with at least 2 elements, such that if \(a < b < c\) and \(a,c \in I\), then \(b \in I\). See .

We have already seen that any open interval \((a,b)\) (where \(a < b\) of course) must be nonempty. For example, it contains the number \(\frac{a+b}{2}\). An unexpected fact is that from a set-theoretic perspective, all intervals have the same “size,” that is, they all have the same cardinality. For example the map \(f(x) := 2x\) takes the interval \([0,1]\) bijectively to the interval \([0,2]\).

Maybe more interestingly, the function \(f(x) := \tan(x)\) is a bijective map from \((-\nicefrac{\pi}{2},\nicefrac{\pi}{2})\) to \({\mathbb{R}}\). Hence the bounded interval \((-\nicefrac{\pi}{2},\nicefrac{\pi}{2})\) has the same cardinality as \({\mathbb{R}}\). It is not completely straightforward to construct a bijective map from \([0,1]\) to say \((0,1)\), but it is possible.

And do not worry, there does exist a way to measure the “size” of subsets of real numbers that “sees” the difference between \([0,1]\) and \([0,2]\). However, its proper definition requires much more machinery than we have right now.

Let us say more about the cardinality of intervals and hence about the cardinality of \({\mathbb{R}}\). We have seen that there exist irrational numbers, that is \({\mathbb{R}}\setminus {\mathbb{Q}}\) is nonempty. The question is: How many irrational numbers are there? It turns out there are a lot more irrational numbers than rational numbers. We have seen that \({\mathbb{Q}}\) is countable, and we will show that \({\mathbb{R}}\) is uncountable. In fact, the cardinality of \({\mathbb{R}}\) is the same as the cardinality of \({\mathcal{P}}({\mathbb{N}})\), although we will not prove this claim here.

\({\mathbb{R}}\) is uncountable.

We give a modified version of Cantor’s original proof from 1874 as this proof requires the least setup. Normally this proof is stated as a contradiction proof, but a proof by contrapositive is easier to understand.

Let \(X \subset {\mathbb{R}}\) be a countably infinite subset such that for any two real numbers \(a < b\), there is an \(x \in X\) such that \(a < x < b\). Were \({\mathbb{R}}\) countable, then we could take \(X = {\mathbb{R}}\). If we show that \(X\) is necessarily a proper subset, then \(X\) cannot equal \({\mathbb{R}}\), and \({\mathbb{R}}\) must be uncountable.

As \(X\) is countably infinite, there is a bijection from \({\mathbb{N}}\) to \(X\). Consequently, we write \(X\) as a sequence of real numbers \(x_1, x_2, x_3,\ldots\), such that each number in \(X\) is given by \(x_j\) for some \(j \in {\mathbb{N}}\).

Let us inductively construct two sequences of real numbers \(a_1,a_2,a_3,\ldots\) and \(b_1,b_2,b_3,\ldots\). Let \(a_1 := x_1\) and \(b_1 := x_1+1\). Note that \(a_1 < b_1\) and \(x_1 \notin (a_1,b_1)\). For \(k > 1\), suppose \(a_{k-1}\) and \(b_{k-1}\) has been defined. Let us also suppose \((a_{k-1},b_{k-1})\) does not contain any \(x_j\) for any \(j=1,\ldots,k-1\).

- Define \(a_k := x_j\), where \(j\) is the smallest \(j \in {\mathbb{N}}\) such that \(x_j \in (a_{k-1},b_{k-1})\). Such an \(x_j\) exists by our assumption on \(X\).
- Next, define \(b_k := x_j\) where \(j\) is the smallest \(j \in {\mathbb{N}}\) such that \(x_j \in (a_{k},b_{k-1})\).

Notice that \(a_k < b_k\) and \(a_{k-1} < a_k < b_k < b_{k-1}\). Also notice that \((a_{k},b_{k})\) does not contain \(x_k\) and hence does not contain any \(x_j\) for \(j=1,\ldots,k\).

Claim: \(a_j < b_k\) for all \(j\) and \(k\) in \({\mathbb{N}}\). Let us first assume \(j < k\). Then \(a_j < a_{j+1} < \cdots < a_{k-1} < a_k < b_k\). Similarly for \(j > k\). The claim follows.

Let \(A = \{ a_j : j \in {\mathbb{N}}\}\) and \(B = \{ b_j : j \in {\mathbb{N}}\}\). By and the claim above we have \[\sup\, A \leq \inf\, B .\] Define \(y := \sup\, A\). The number \(y\) cannot be a member of \(A\). If \(y = a_j\) for some \(j\), then \(y < a_{j+1}\), which is impossible. Similarly \(y\) cannot be a member of \(B\). Therefore, \(a_j < y\) for all \(j\in {\mathbb{N}}\) and \(y < b_j\) for all \(j\in {\mathbb{N}}\). In other words \(y \in (a_j,b_j)\) for all \(j\in {\mathbb{N}}\).

Finally we must show that \(y \notin X\). If we do so, then we will have constructed a real number not in \(X\) showing that \(X\) must have been a proper subset. Take any \(x_k \in X\). By the above construction \(x_k \notin (a_k,b_k)\), so \(x_k \not= y\) as \(y \in (a_k,b_k)\).

Therefore, the sequence \(x_1,x_2,\ldots\) cannot contain all elements of \({\mathbb{R}}\) and thus \({\mathbb{R}}\) is uncountable.

## Exercises

For \(a < b\), construct an explicit bijection from \((a,b]\) to \((0,1]\).

Suppose \(f \colon [0,1] \to (0,1)\) is a bijection. Using \(f\), construct a bijection from \([-1,1]\) to \({\mathbb{R}}\).

[exercise:intervaldef] Suppose \(I \subset {\mathbb{R}}\) is a subset with at least 2 elements such that if \(a < b < c\) and \(a, c \in I\), then it is one of the nine types of intervals explicitly given in this section. Furthermore, prove that the intervals given in this section all satisfy this property.

Construct an explicit bijection from \((0,1]\) to \((0,1)\). Hint: One approach is as follows: First map \((\nicefrac{1}{2},1]\) to \((0,\nicefrac{1}{2}]\), then map \((\nicefrac{1}{4},\nicefrac{1}{2}]\) to \((\nicefrac{1}{2},\nicefrac{3}{4}]\), etc…. Write down the map explicitly, that is, write down an algorithm that tells you exactly what number goes where. Then prove that the map is a bijection.

Construct an explicit bijection from \([0,1]\) to \((0,1)\).

a) Show that every closed interval \([a,b]\) is the intersection of countably many open intervals. b) Show that every open interval \((a,b)\) is a countable union of closed intervals. c) Show that an intersection of a possibly infinite family of closed intervals is either empty, a single point, or a closed interval.

Suppose \(S\) is a set of disjoint open intervals in \({\mathbb{R}}\). That is, if \((a,b) \in S\) and \((c,d) \in S\), then either \((a,b) = (c,d)\) or \((a,b) \cap (c,d) = \emptyset\). Prove \(S\) is a countable set.

Prove that the cardinality of \([0,1]\) is the same as the cardinality of \((0,1)\) by showing that \(\left\lvert {[0,1]} \right\rvert \leq \left\lvert {(0,1)} \right\rvert\) and \(\left\lvert {(0,1)} \right\rvert \leq \left\lvert {[0,1]} \right\rvert\). See . Note that this requires the Cantor-Bernstein-Schroeder theorem we stated without proof. Also note that this proof does not give you an explicit bijection.

A number \(x\) is *algebraic* if \(x\) is a root of a polynomial with integer coefficients, in other words, \(a_n x^n + a_{n-1} x^{n-1} + \ldots + a_1 x + a_0 = 0\) where all \(a_n \in {\mathbb{Z}}\). a) Show that there are only countably many algebraic numbers. b) Show that there exist non-algebraic numbers (follow in the footsteps of Cantor, use uncountability of \({\mathbb{R}}\)). Hint: Feel free to use the fact that a polynomial of degree \(n\) has at most \(n\) real roots.

### Decimal representation of the reals

*Note: 1 lecture (optional)*

We often think of real numbers as their *decimal representation*. For a positive integer \(n\), we find the digits \(d_K,d_{K-1},\ldots,d_2,d_1,d_0\) for some \(K\), where each \(d_j\) is an integer between \(0\) and \(9\), then \[n = d_K {10}^K + d_{K-1} {10}^{K-1} + \cdots + d_2 {10}^2 + d_1 10 + d_0 .\] We often assume \(d_K \not= 0\). To represent \(n\) we write the sequence of digits: \(n = d_K d_{K-1} \cdots d_2 d_1 d_0\). By a (decimal) *digit*, we mean an integer between \(0\) and \(9\).

Similarly we represent some rational numbers. That is, for certain numbers \(x\), we can find negative integer \(-M\), a positive integer \(K\), and digits \(d_K,d_{K-1},\ldots,d_1,d_0,d_{-1},\ldots,d_{-M}\), such that \[x = d_K {10}^K + d_{K-1} {10}^{K-1} + \cdots + d_2 {10}^2 + d_1 10 + d_0 + d_{-1} {10}^{-1} + d_{-2} {10}^{-2} + \cdots + d_{-M} {10}^{-M} .\] We write \(x = d_K d_{K-1} \cdots d_1 d_0 \, . \, d_{-1} d_{-2} \cdots d_{-M}\).

Not every real number has such a representation, even the simple rational number \(\nicefrac{1}{3}\) does not. The irrational number \(\sqrt{2}\) does not have such a representation either. To get a representation for all real numbers we must allow infinitely many digits.

Let us from now on consider only real numbers in the interval \((0,1]\). If we find a representation for these, we simply add integers to them to obtain a representation for all real numbers. Suppose we take an infinite sequence of decimal digits: \[0.d_1d_2d_3\ldots.\] That is, we have a digit \(d_j\) for every \(j \in {\mathbb{N}}\). We have renumbered the digits to avoid the negative signs. We say this sequence of digits represents a real number \(x\) if \[x = \sup_{n \in {\mathbb{N}}} \left( \frac{d_1}{10} + \frac{d_2}{{10}^2} + \frac{d_3}{{10}^3} + \cdots + \frac{d_n}{{10}^n} \right) .\] We call \[D_n := \frac{d_1}{10} + \frac{d_2}{{10}^2} + \frac{d_3}{{10}^3} + \cdots + \frac{d_n}{{10}^n}\] the truncation of \(x\) to \(n\) decimal digits.

[prop:decimalprop]

- Every infinite sequence of digits \(0.d_1d_2d_3\ldots\) represents a unique real number \(x \in [0,1]\).
- For every \(x \in (0,1]\) there exists an infinite sequence of digits \(0.d_1d_2d_3\ldots\) that represents \(x\). There exists a unique representation such that \[D_n < x \leq D_n+\frac{1}{{10}^n} \qquad \text{for all $n \in {\mathbb{N}}$}.\]

Let us start with the first item. Suppose there is an infinite sequence of digits \(0.d_1d_2d_3\ldots\). We use the geometric sum formula to write \[\begin{split} D_n = \frac{d_1}{10} + \frac{d_2}

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[7]/div/p[7]/span, line 1, column 3 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

We move on to the second item. Take \(x \in (0,1]\). First let us tackle the existence. By convention define \(D_0 := 0\), then automatically we obtain \(D_0 < x \leq D_0 + {10}^{-0}\). Suppose for induction that we defined all the digits \(d_1,d_2,\ldots,d_n\), and that \(D_n < x \leq D_n + {10}^{-n}\). We need to define \(d_{n+1}\).

By the of the real numbers we find an integer \(j\) such that \(x-D_n \leq j {10}^{-(n+1)}\). We take the least such \(j\) and obtain \[\label{eq:theDnjineq} (j-1){10}^{-(n+1)} < x-D_n \leq j {10}^{-(n+1)} .\] Let \(d_{n+1} := j-1\). As \(D_n < x\), then \(d_{n+1} = j-1 \geq 0\). On the other hand since \(x-D_n \leq {10}^{-n}\) we have that \(j\) is at most 10, and therefore \(d_{n+1} \leq 9\). So \(d_{n+1}\) is a decimal digit. Since \(D_{n+1} = D_n + d_{n+1} {10}^{-(n+1)}\) we add \(D_n\) to the inequality [eq:theDnjineq] above: \[D_{n+1} = D_n + (j-1){10}^{-(n+1)} < x \leq %D_n + j {10}^{-(n+1)} = D_n + (j-1) {10}^{-(n+1)} + {10}^{-(n+1)} = D_{n+1} + {10}^{-(n+1)} .\] And so \(D_{n+1} < x \leq D_{n+1} + {10}^{-(n+1)}\) holds. We have inductively defined an infinite sequence of digits \(0.d_1d_2d_3\ldots\). As \(D_n < x\) for all \(n\), then \(\sup \{ D_n : n \in {\mathbb{N}}\} \leq x\). As \(x-{10}^{-n} \leq D_n\), then \(x-{10}^{-n} \leq \sup \{ D_m : m \in {\mathbb{N}}\}\) for all \(n\). The two inequalities together imply \(\sup \{ D_n : n \in {\mathbb{N}}\} = x\).

What is left to show is the uniqueness. Suppose \(0.e_1e_2e_3\ldots\) is another representation of \(x\). Let \(E_n\) be the \(n\)-digit truncation of \(0.e_1e_2e_3\ldots\), and suppose \(E_n < x \leq E_n + {10}^{-n}\) for all \(n \in {\mathbb{N}}\). Suppose for some \(K \in {\mathbb{N}}\), \(e_n = d_n\) for all \(n < K\), so \(D_{K-1} = E_{K-1}\). Then \[E_K = D_{K-1} + e_K{10}^{-K} < x \leq E_K + {10}^{-K} = D_{K-1} + e_K{10}^{-K} + {10}^{-K} .\] Subtracting \(D_{K-1}\) and multiplying by \({10}^{K}\) we get \[e_K < (x - D_{K-1}){10}^K \leq e_K + 1 .\] Similarly we obtain \[d_K < (x - D_{K-1}){10}^K \leq d_K + 1 .\] Hence, both \(e_K\) and \(d_K\) are the largest integer \(j\) such that \(j < (x - D_{K-1}){10}^K\), and therefore \(e_K = d_K\). That is, the representation is unique.

The representation is not unique if we do not require the extra condition in the proposition. For example, for the number \(\nicefrac{1}{2}\) the method in the proof obtains the representation \[0.49999\ldots .\] However, we also have the representation \(0.5000\ldots\). The key requirement that makes the representation unique is \(D_n < x\) for all \(n\). The inequality \(x \leq D_n + {10}^{-n}\) is true for every representation by the computation in the beginning of the proof.

The only numbers that have nonunique representations are ones that end either in an infinite sequence of \(0\)s or \(9\)s, because the only representation for which \(D_n = x\) is one where all digits past \(n\)th one are zero. In this case there are exactly two representations of \(x\) (see the exercises).

Let us give another proof of the uncountability of the reals using decimal representations. This is Cantor’s second proof, and is probably more well known. While this proof may seem shorter, it is because we have already done the hard part above and we are left with a slick trick to prove that \({\mathbb{R}}\) is uncountable. This trick is called *Cantor diagonalization* and finds use in other proofs as well.

The set \((0,1]\) is uncountable.

Let \(X := \{ x_1,x_2,x_3,\ldots \}\) be any countable subset of real numbers in \((0,1]\). We will construct a real number not in \(X\). Let \[x_n = 0.d_1^nd_2^nd_3^n\ldots\] be the unique representation from the proposition, that is \(d_j^n\) is the \(j\)th digit of the \(n\)th number. Let \(e_n := 1\) if \(d_n^n \not= 1\), and let \(e_n := 2\) if \(d_n^n = 1\). Let \(E_n\) be the \(n\)-digit truncation of \(y = 0.e_1e_2e_3\ldots\). Because all the digits are nonzero we get that \(E_n < E_{n+1} \leq y\). Therefore \[E_n < y \leq E_n + {10}^{-n}\] for all \(n\), and the representation is the unique one for \(y\) from the proposition. But for every \(n\), the \(n\)th digit of \(y\) is different from the \(n\)th digit of \(x_n\), so \(y \not= x_n\). Therefore \(y \notin X\), and as \(X\) was an arbitrary countable subset, \((0,1]\) must be uncountable.

Using decimal digits we can also find lots of numbers that are not rational. The following proposition is true for every rational number, but we give it only for \(x \in (0,1]\) for simplicity.

[prop:rationaldecimal] If \(x \in (0,1]\) is a rational number and \(x = 0.d_1d_2d_3\ldots\), then the decimal digits eventually start repeating. That is, there are positive integers \(N\) and \(P\), such that for all \(n \geq N\), \(d_n = d_{n+P}\).

Let \(x = \nicefrac{p}{q}\) for positive integers \(p\) and \(q\). Let us suppose \(x\) is a number with a unique representation, as otherwise we have seen above that both its representations are repeating.

To compute the first digit we take \(10 p\) and divide by \(q\). The quotient is the first digit \(d_1\) and the remainder \(r\) is some integer between 0 and \(q-1\). That is, \(d_1\) is the largest integer such that \(d_1 q \leq 10p\) and then \(r = 10p - d_1q\).

The next digit is computed by dividing \(10 r\) by \(q\), and so on. We notice that at each step there are at most \(q\) possible remainders and hence at some point the process must start repeating. In fact we see that \(P\) is at most \(q\).

The converse of the proposition is also true and is left as an exercise.

The number \[x = 0.101001000100001000001\ldots,\] is irrational. That is, the digits are \(n\) zeros, then a one, then \(n+1\) zeros, then a one, and so on and so forth. The fact that \(x\) is irrational follows from the proposition; the digits never start repeating. For every \(P\), if we go far enough, we find a 1 that is followed by at least \(P+1\) zeros.

## Exercises

What is the decimal representation of \(1\) guaranteed by ? Make sure to show that it does satisfy the condition.

Prove the converse of , that is, if the digits in the decimal representation of \(x\) are eventually repeating, then \(x\) must be rational.

Show that real numbers \(x \in (0,1)\) with nonunique decimal representation are exactly the rational numbers that can be written as \(\frac{m}{10^n}\) for some integers \(m\) and \(n\). In this case show that there exist exactly two representations of \(x\).

Let \(b \geq 2\) be an integer. Define a representation of a real number in \([0,1]\) in terms of base \(b\) rather than base 10 and prove for base \(b\).

Using the previous exercise with \(b=2\) (binary), show that cardinality of \({\mathbb{R}}\) is the same as the cardinality of \({\mathcal{P}}({\mathbb{N}})\), obtaining yet another (though related) proof that \({\mathbb{R}}\) is uncountable. Hint: Construct two injections, one from \([0,1]\) to \({\mathcal{P}}({\mathbb{N}})\) and one from \({\mathcal{P}}({\mathbb{N}})\) to \([0,1]\). Hint 2: Given a set \(A \subset {\mathbb{N}}\), let the \(n\)th binary digit of \(x\) be 1 if \(n\in A\).

Construct a bijection between \([0,1]\) and \([0,1] \times [0,1]\). Hint: consider even and odd digits, and be careful about the uniqueness of representation.

## Sequences and Series

### Sequences and limits

*Note: 2.5 lectures*

Analysis is essentially about taking limits. The most basic type of a limit is a limit of a sequence of real numbers. We have already seen sequences used informally. Let us give the formal definition.

A *sequence* (of real numbers) is a function \(x \colon {\mathbb{N}}\to {\mathbb{R}}\). Instead of \(x(n)\) we usually denote the \(n\)th element in the sequence by \(x_n\). We use the notation \(\{ x_n \}\), or more precisely \[\{ x_n \}_{n=1}^\infty,\] to denote a sequence.

A sequence \(\{ x_n \}\) is *bounded* if there exists a \(B \in {\mathbb{R}}\) such that \[\left\lvert {x_n} \right\rvert \leq B \qquad \text{for all $n \in {\mathbb{N}}$.}\] In other words, the sequence \(\{x_n\}\) is bounded whenever the set \(\{ x_n : n \in {\mathbb{N}}\}\) is bounded.

When we need to give a concrete sequence we often give each term as a formula in terms of \(n\). For example, \(\{ \nicefrac{1}{n} \}_{n=1}^\infty\), or simply \(\{ \nicefrac{1}{n} \}\), stands for the sequence \(1, \nicefrac{1}{2}, \nicefrac{1}{3}, \nicefrac{1}{4}, \nicefrac{1}{5}, \ldots\). The sequence \(\{ \nicefrac{1}{n} \}\) is a bounded sequence (\(B=1\) will suffice). On the other hand the sequence \(\{ n \}\) stands for \(1,2,3,4,\ldots\), and this sequence is not bounded (why?).

While the notation for a sequence is similar^{7} to that of a set, the notions are distinct. For example, the sequence \(\{ {(-1)}^n \}\) is the sequence \(-1,1,-1,1,-1,1,\ldots\), whereas the set of values, the *range of the sequence*, is just the set \(\{ -1, 1 \}\). We can write this set as \(\{ {(-1)}^n : n \in {\mathbb{N}}\}\). When ambiguity can arise, we use the words *sequence* or *set* to distinguish the two concepts.

Another example of a sequence is the so-called *constant sequence*. That is a sequence \(\{ c \} = c,c,c,c,\ldots\) consisting of a single constant \(c \in {\mathbb{R}}\) repeating indefinitely.

We now get to the idea of a *limit of a sequence*. We will see in that the notation below is well defined. That is, if a limit exists, then it is unique. So it makes sense to talk about *the* limit of a sequence.

A sequence \(\{ x_n \}\) is said to *converge* to a number \(x \in {\mathbb{R}}\), if for every \(\epsilon > 0\), there exists an \(M \in {\mathbb{N}}\) such that \(\left\lvert {x_n - x} \right\rvert < \epsilon\) for all \(n \geq M\). The number \(x\) is said to be the *limit* of \(\{ x_n \}\). We write \[\lim_{n\to \infty} x_n := x .\]

A sequence that converges is said to be *convergent*. Otherwise, the sequence is said to be *divergent*.

It is good to know intuitively what a limit means. It means that eventually every number in the sequence is close to the number \(x\). More precisely, we can get arbitrarily close to the limit, provided we go far enough in the sequence. It does not mean we ever reach the limit. It is possible, and quite common, that there is no \(x_n\) in the sequence that equals the limit \(x\). We illustrate the concept in . In the figure we first think of the sequence as a graph, as it is a function of \({\mathbb{N}}\). Secondly we also plot it as a sequence of labeled points on the real line.

When we write \(\lim\, x_n = x\) for some real number \(x\), we are saying two things. First, that \(\{ x_n \}\) is convergent, and second that the limit is \(x\).

The above definition is one of the most important definitions in analysis, and it is necessary to understand it perfectly. The key point in the definition is that given *any* \(\epsilon > 0\), we can find an \(M\). The \(M\) can depend on \(\epsilon\), so we only pick an \(M\) once we know \(\epsilon\). Let us illustrate this concept on a few examples.

The constant sequence \(1,1,1,1,\ldots\) is convergent and the limit is 1. For every \(\epsilon > 0\), we pick \(M = 1\).

Claim: The sequence \(\{ \nicefrac{1}{n} \}\) is convergent and \[\lim_{n\to \infty} \frac{1}{n} = 0 .\] Proof: Given an \(\epsilon > 0\), we find an \(M \in {\mathbb{N}}\) such that \(0 < \nicefrac{1}{M} < \epsilon\) ( at work). Then for all \(n \geq M\) we have that \[\left\lvert {x_n - 0} \right\rvert = \left\lvert {\frac{1}{n}} \right\rvert = \frac{1}{n} \leq \frac{1}{M} < \epsilon .\]

The sequence \(\{ {(-1)}^n \}\) is divergent. Proof: If there were a limit \(x\), then for \(\epsilon = \frac{1}{2}\) we expect an \(M\) that satisfies the definition. Suppose such an \(M\) exists, then for an even \(n \geq M\) we compute \[\nicefrac{1}{2} > \left\lvert {x_n - x} \right\rvert = \left\lvert {1 - x} \right\rvert \qquad \text{and} \qquad \nicefrac{1}{2} > \left\lvert {x_{n+1} - x} \right\rvert = \left\lvert {-1 - x} \right\rvert .\] But \[2 = \left\lvert {1 - x - (-1 -x)} \right\rvert \leq \left\lvert {1 - x} \right\rvert + \left\lvert {-1 -x} \right\rvert < \nicefrac{1}{2} + \nicefrac{1}{2} = 1 ,\] and that is a contradiction.

[prop:limisunique] A convergent sequence has a unique limit.

The proof of this proposition exhibits a useful technique in analysis. Many proofs follow the same general scheme. We want to show a certain quantity is zero. We write the quantity using the triangle inequality as two quantities, and we estimate each one by arbitrarily small numbers.

Suppose the sequence \(\{ x_n \}\) has the limit \(x\) and the limit \(y\). Take an arbitrary \(\epsilon > 0\). From the definition find an \(M_1\) such that for all \(n \geq M_1\), \(\left\lvert {x_n-x} \right\rvert < \nicefrac{\epsilon}{2}\). Similarly find an \(M_2\) such that for all \(n \geq M_2\) we have \(\left\lvert {x_n-y} \right\rvert < \nicefrac{\epsilon}{2}\). Take \(M := \max \{M_1, M_2\}\). For \(n \geq M\) (so that both \(n \geq M_1\) and \(n \geq M_2\)) we have \[\begin{split} \left\lvert {y-x} \right\rvert & = \left\lvert {x_n-x - (x_n -y)} \right\rvert \\ & \leq \left\lvert {x_n-x} \right\rvert + \left\lvert {x_n -y} \right\rvert \\ & < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon . \end{split}\] As \(\left\lvert {y-x} \right\rvert < \epsilon\) for all \(\epsilon > 0\), then \(\left\lvert {y-x} \right\rvert = 0\) and \(y=x\). Hence the limit (if it exists) is unique.

A convergent sequence \(\{ x_n \}\) is bounded.

Suppose \(\{ x_n \}\) converges to \(x\). Thus there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \(\left\lvert {x_n - x} \right\rvert < 1\). Let \(B_1 := \left\lvert {x} \right\rvert+1\) and note that for \(n \geq M\) we have \[\begin{split} \left\lvert {x_n} \right\rvert & = \left\lvert {x_n - x + x} \right\rvert \\ & \leq \left\lvert {x_n - x} \right\rvert + \left\lvert {x} \right\rvert \\ & < 1 + \left\lvert {x} \right\rvert = B_1 . \end{split}\] The set \(\{ \left\lvert {x_1} \right\rvert, \left\lvert {x_2} \right\rvert, \ldots, \left\lvert {x_{M-1}} \right\rvert \}\) is a finite set and hence let \[B_2 := \max \{ \left\lvert {x_1} \right\rvert, \left\lvert {x_2} \right\rvert, \ldots, \left\lvert {x_{M-1}} \right\rvert \} .\] Let \(B := \max \{ B_1, B_2 \}\). Then for all \(n \in {\mathbb{N}}\) we have \[\left\lvert {x_n} \right\rvert \leq B. \qedhere\]

The sequence \(\{ {(-1)}^n \}\) shows that the converse does not hold. A bounded sequence is not necessarily convergent.

Let us show the sequence \(\left\{ \frac{n^2+1}{n^2+n} \right\}\) converges and \[\lim_{n\to\infty} \frac{n^2+1}{n^2+n} = 1 .\]

Given \(\epsilon > 0\), find \(M \in {\mathbb{N}}\) such that \(\frac{1}{M+1} < \epsilon\). Then for any \(n \geq M\) we have \[\begin{split} %\abs{\frac{n^2+1}{n^2+n} - 1} & = %\abs{\frac{n^2+1 - (n^2+n)}{n^2+n}} \\ \left\lvert {\frac{n^2+1}{n^2+n} - 1} \right\rvert = \left\lvert {\frac{n^2+1 - (n^2+n)}{n^2+n}} \right\rvert & = \left\lvert {\frac{1 - n}{n^2+n}} \right\rvert \\ & = \frac{n-1}{n^2+n} \\ & \leq \frac{n}{n^2+n} = \frac{1}{n+1} \\ & \leq \frac{1}{M+1} < \epsilon . \end{split}\] Therefore, \(\lim \frac{n^2+1}{n^2+n} = 1\).

#### Monotone sequences

The simplest type of a sequence is a monotone sequence. Checking that a monotone sequence converges is as easy as checking that it is bounded. It is also easy to find the limit for a convergent monotone sequence, provided we can find the supremum or infimum of a countable set of numbers.

A sequence \(\{ x_n \}\) is *monotone increasing* if \(x_n \leq x_{n+1}\) for all \(n \in {\mathbb{N}}\). A sequence \(\{ x_n \}\) is *monotone decreasing* if \(x_n \geq x_{n+1}\) for all \(n \in {\mathbb{N}}\). If a sequence is either monotone increasing or monotone decreasing, we can simply say the sequence is *monotone*. Some authors also use the word *monotonic*.

For example, \(\{ \nicefrac{1}{n} \}\) is monotone decreasing, the constant sequence \(\{ 1 \}\) is both monotone increasing and monotone decreasing, and \(\{ {(-1)}^n \}\) is not monotone. First few terms of a sample monotone increasing sequence are shown in .

[thm:monotoneconv] A monotone sequence \(\{ x_n \}\) is bounded if and only if it is convergent.

Furthermore, if \(\{ x_n \}\) is monotone increasing and bounded, then \[\lim_{n\to \infty} x_n = \sup \{ x_n : n \in {\mathbb{N}}\} .\] If \(\{ x_n \}\) is monotone decreasing and bounded, then \[\lim_{n\to \infty} x_n = \inf \{ x_n : n \in {\mathbb{N}}\} .\]

Let us suppose the sequence is monotone increasing. Suppose the sequence is bounded, so there exists a \(B\) such that \(x_n \leq B\) for all \(n\), that is the set \(\{ x_n : n \in {\mathbb{N}}\}\) is bounded from above. Let \[x := \sup \{ x_n : n \in {\mathbb{N}}\} .\] Let \(\epsilon > 0\) be arbitrary. As \(x\) is the supremum, then there must be at least one \(M \in {\mathbb{N}}\) such that \(x_{M} > x-\epsilon\) (because \(x\) is the supremum). As \(\{ x_n \}\) is monotone increasing, then it is easy to see (by ) that \(x_n \geq x_{M}\) for all \(n \geq M\). Hence \[\left\lvert {x_n-x} \right\rvert = x-x_n \leq x-x_{M} < \epsilon .\] Therefore the sequence converges to \(x\). We already know that a convergent sequence is bounded, which completes the other direction of the implication.

The proof for monotone decreasing sequences is left as an exercise.

Take the sequence \(\{ \frac{1}{\sqrt{n}} \}\).

First \(\frac{1}{\sqrt{n}} > 0\) for all \(n \in {\mathbb{N}}\), and hence the sequence is bounded from below. Let us show that it is monotone decreasing. We start with \(\sqrt{n+1} \geq \sqrt{n}\) (why is that true?). From this inequality we obtain \[\frac{1}{\sqrt{n+1}} \leq \frac{1}{\sqrt{n}} .\] So the sequence is monotone decreasing and bounded from below (hence bounded). We apply the theorem to note that the sequence is convergent and in fact \[\lim_{n\to \infty} \frac{1}{\sqrt{n}} = \inf \left\{ \frac{1}{\sqrt{n}} : n \in {\mathbb{N}}\right\} .\] We already know that the infimum is greater than or equal to 0, as 0 is a lower bound. Take a number \(b \geq 0\) such that \(b \leq \frac{1}{\sqrt{n}}\) for all \(n\). We square both sides to obtain \[b^2 \leq \frac{1}{n} \qquad \text{for all $n \in {\mathbb{N}}$}.\] We have seen before that this implies that \(b^2 \leq 0\) (a consequence of the ). As we also have \(b^2 \geq 0\), then \(b^2 = 0\) and so \(b = 0\). Hence \(b=0\) is the greatest lower bound, and \(\lim \frac{1}{\sqrt{n}} = 0\).

A word of caution: We must show that a monotone sequence is bounded in order to use . For example, the sequence \(\{ 1 + \nicefrac{1}{2} + \cdots + \nicefrac{1}{n} \}\) is a monotone increasing sequence that grows very slowly. We will see, once we get to series, that this sequence has no upper bound and so does not converge. It is not at all obvious that this sequence has no upper bound.

A common example of where monotone sequences arise is the following proposition. The proof is left as an exercise.

[prop:supinfseq] Let \(S \subset {\mathbb{R}}\) be a nonempty bounded set. Then there exist monotone sequences \(\{ x_n \}\) and \(\{ y_n \}\) such that \(x_n, y_n \in S\) and \[\sup\,S = \lim_{n\to \infty} x_n \qquad \text{and} \qquad \inf\,S = \lim_{n\to\infty} y_n .\]

#### Tail of a sequence

For a sequence \(\{ x_n \}\), the *\(K\)-tail* (where \(K \in {\mathbb{N}}\)) or just the *tail* of the sequence is the sequence starting at \(K+1\), usually written as \[\{ x_{n+K} \}_{n=1}^\infty \qquad \text{or} \qquad \{ x_n \}_{n=K+1}^\infty .\]

The main result about the tail of a sequence is the following proposition.

Let \(\{ x_n \}_{n=1}^\infty\) be a sequence. Then the following statements are equivalent:

- [prop:ktail:i] The sequence \(\{ x_n \}_{n=1}^\infty\) converges.
- [prop:ktail:ii] The \(K\)-tail \(\{ x_{n+K} \}_{n=1}^\infty\) converges for all \(K \in {\mathbb{N}}\).
- [prop:ktail:iii] The \(K\)-tail \(\{ x_{n+K} \}_{n=1}^\infty\) converges for some \(K \in {\mathbb{N}}\).

Furthermore, if any (and hence all) of the limits exist, then for any \(K \in {\mathbb{N}}\) \[\lim_{n\to \infty} x_n = \lim_{n \to \infty} x_{n+K} .\]

It is clear that [prop:ktail:ii] implies [prop:ktail:iii]. We will therefore show first that [prop:ktail:i] implies [prop:ktail:ii], and then we will show that [prop:ktail:iii] implies [prop:ktail:i]. In the process we will also show that the limits are equal.

Let us start with [prop:ktail:i] implies [prop:ktail:ii]. Suppose \(\{x_n \}\) converges to some \(x \in {\mathbb{R}}\). Let \(K \in {\mathbb{N}}\) be arbitrary. Define \(y_n := x_{n+K}\), we wish to show that \(\{ y_n \}\) converges to \(x\). That is, given an \(\epsilon > 0\), there exists an \(M \in {\mathbb{N}}\) such that \(\left\lvert {x-x_n} \right\rvert < \epsilon\) for all \(n \geq M\). Note that \(n \geq M\) implies \(n+K \geq M\). Therefore, it is true that for all \(n \geq M\) we have that \[\left\lvert {x-y_n} \right\rvert = \left\lvert {x-x_{n+K}} \right\rvert < \epsilon .\] Therefore \(\{ y_n \}\) converges to \(x\).

Let us move to [prop:ktail:iii] implies [prop:ktail:i]. Let \(K \in {\mathbb{N}}\) be given, define \(y_n := x_{n+K}\), and suppose that \(\{ y_n \}\) converges \(x \in {\mathbb{R}}\). That is, given an \(\epsilon > 0\), there exists an \(M' \in {\mathbb{N}}\) such that \(\left\lvert {x-y_n} \right\rvert < \epsilon\) for all \(n \geq M'\). Let \(M := M'+K\). Then \(n \geq M\) implies \(n-K \geq M'\). Thus, whenever \(n \geq M\) we have \[\left\lvert {x-x_n} \right\rvert = \left\lvert {x-y_{n-K}} \right\rvert < \epsilon.\] Therefore \(\{ x_n \}\) converges to \(x\).

Essentially, the limit does not care about how the sequence begins, it only cares about the tail of the sequence. That is, the beginning of the sequence may be arbitrary.

For example, the sequence defined by \(x_n := \frac{n}{n^2+16}\) is decreasing if we start at \(n=4\) (it is increasing before). That is, \(\{ x_n \} = \nicefrac{1}{17}, \nicefrac{1}{10}, \nicefrac{3}{25}, \nicefrac{1}{8}, \nicefrac{5}{41}, \nicefrac{3}{26}, \nicefrac{7}{65}, \nicefrac{1}{10}, \nicefrac{9}{97}, \nicefrac{5}{58},\ldots\), and \[\nicefrac{1}{17} < \nicefrac{1}{10} < \nicefrac{3}{25} < \nicefrac{1}{8} > \nicefrac{5}{41} > \nicefrac{3}{26} > \nicefrac{7}{65} > \nicefrac{1}{10} > \nicefrac{9}{97} > \nicefrac{5}{58} > \ldots .\] That is if we throw away the first 3 terms and look at the 3 tail it is decreasing. The proof is left as an exercise. Since the 3-tail is monotone and bounded below by zero, it is convergent, and therefore the sequence is convergent.

#### Subsequences

A very useful concept related to sequences is that of a subsequence. A subsequence of \(\{ x_n \}\) is a sequence that contains only some of the numbers from \(\{ x_n \}\) in the same order.

Let \(\{ x_n \}\) be a sequence. Let \(\{ n_i \}\) be a strictly increasing sequence of natural numbers (that is \(n_1 < n_2 < n_3 < \cdots\)). The sequence \[\{ x_{n_i} \}_{i=1}^\infty\] is called a *subsequence* of \(\{ x_n \}\).

For example, take the sequence \(\{ \nicefrac{1}{n} \}\). The sequence \(\{ \nicefrac{1}{3n} \}\) is a subsequence. To see how these two sequences fit in the definition, take \(n_i := 3i\). The numbers in the subsequence must come from the original sequence, so \(1,0,\nicefrac{1}{3},0, \nicefrac{1}{5},\ldots\) is not a subsequence of \(\{ \nicefrac{1}{n} \}\). Similarly order must be preserved, so the sequence \(1,\nicefrac{1}{3},\nicefrac{1}{2},\nicefrac{1}{5},\ldots\) is not a subsequence of \(\{ \nicefrac{1}{n} \}\).

A tail of a sequence is one special type of a subsequence. For an arbitrary subsequence, we have the following proposition about convergence.

[prop:seqtosubseq] If \(\{ x_n \}\) is a convergent sequence, then any subsequence \(\{ x_{n_i} \}\) is also convergent and \[\lim_{n\to \infty} x_n = \lim_{i\to \infty} x_{n_i} .\]

Suppose \(\lim_{n\to \infty} x_n = x\). That means that for every \(\epsilon > 0\) we have an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) \[\left\lvert {x_n - x} \right\rvert < \epsilon .\] It is not hard to prove (do it!) by that \(n_i \geq i\). Hence \(i \geq M\) implies \(n_i \geq M\). Thus, for all \(i \geq M\) we have \[\left\lvert {x_{n_i} - x} \right\rvert < \epsilon ,\] and we are done.

Existence of a convergent subsequence does not imply convergence of the sequence itself. Take the sequence \(0,1,0,1,0,1,\ldots\). That is, \(x_n = 0\) if \(n\) is odd, and \(x_n = 1\) if \(n\) is even. The sequence \(\{ x_n \}\) is divergent, however, the subsequence \(\{ x_{2n} \}\) converges to 1 and the subsequence \(\{ x_{2n+1} \}\) converges to 0. Compare .

## Exercises

In the following exercises, feel free to use what you know from calculus to find the limit, if it exists. But you must *prove* that you found the correct limit, or prove that the series is divergent.

Is the sequence \(\{ 3n \}\) bounded? Prove or disprove.

Is the sequence \(\{ n \}\) convergent? If so, what is the limit?

Is the sequence \(\left\{ \dfrac{{(-1)}^n}{2n} \right\}\) convergent? If so, what is the limit?

Is the sequence \(\{ 2^{-n} \}\) convergent? If so, what is the limit?

Is the sequence \(\left\{ \dfrac{n}{n+1} \right\}\) convergent? If so, what is the limit?

Is the sequence \(\left\{ \dfrac{n}{n^2+1} \right\}\) convergent? If so, what is the limit?

[exercise:absconv] Let \(\{ x_n \}\) be a sequence.

- Show that \(\lim\, x_n = 0\) (that is, the limit exists and is zero) if and only if \(\lim \left\lvert {x_n} \right\rvert = 0\).
- Find an example such that \(\{ \left\lvert {x_n} \right\rvert \}\) converges and \(\{ x_n \}\) diverges.

Is the sequence \(\left\{ \dfrac{2^n}{n!} \right\}\) convergent? If so, what is the limit?

Show that the sequence \(\left\{ \dfrac{1}{\sqrt[3]{n}} \right\}\) is monotone, bounded, and use to find the limit.

Show that the sequence \(\left\{ \dfrac{n+1}{n} \right\}\) is monotone, bounded, and use to find the limit.

Finish the proof of for monotone decreasing sequences.

Prove .

Let \(\{ x_n \}\) be a convergent monotone sequence. Suppose there exists a \(k \in {\mathbb{N}}\) such that \[\lim_{n\to \infty} x_n = x_k .\] Show that \(x_n = x_k\) for all \(n \geq k\).

Find a convergent subsequence of the sequence \(\{ {(-1)}^n \}\).

Let \(\{x_n\}\) be a sequence defined by \[x_n := \begin{cases} n & \text{if $n$ is odd} , \\ \nicefrac{1}{n} & \text{if $n$ is even} . \end{cases}\]

- Is the sequence bounded? (prove or disprove)
- Is there a convergent subsequence? If so, find it.

Let \(\{ x_n \}\) be a sequence. Suppose there are two convergent subsequences \(\{ x_{n_i} \}\) and \(\{ x_{m_i} \}\). Suppose \[\lim_{i\to\infty} x_{n_i} = a \qquad \text{and} \qquad \lim_{i\to\infty} x_{m_i} = b,\] where \(a \not= b\). Prove that \(\{ x_n \}\) is not convergent, without using .

Find a sequence \(\{ x_n \}\) such that for any \(y \in {\mathbb{R}}\), there exists a subsequence \(\{ x_{n_i} \}\) converging to \(y\).

Let \(\{ x_n \}\) be a sequence and \(x \in {\mathbb{R}}\). Suppose for any \(\epsilon > 0\), there is an \(M\) such that for all \(n \geq M\), \(\left\lvert {x_n-x} \right\rvert \leq \epsilon\). Show that \(\lim\, x_n = x\).

Let \(\{ x_n \}\) be a sequence and \(x \in {\mathbb{R}}\) such that there exists a \(k \in {\mathbb{N}}\) such that for all \(n \geq k\), \(x_n = x\). Prove that \(\{ x_n \}\) converges to \(x\).

Let \(\{ x_n \}\) be a sequence and define a sequence \(\{ y_n \}\) by \(y_{2k} := x_{k^2}\) and \(y_{2k-1} = x_k\) for all \(k \in {\mathbb{N}}\). Prove that \(\{ x_n \}\) converges if and only if \(\{ y_n \}\) converges. Furthermore, prove that if they converge, then \(\lim\, x_n = \lim\, y_n\).

Show that the 3-tail of the sequence defined by \(x_n := \frac{n}{n^2+16}\) is monotone decreasing. Hint: Suppose \(n \geq m \geq 4\) and consider the numerator of the expression \(x_n-x_m\).

Suppose that \(\{ x_n \}\) is a sequence such that the subsequences \(\{ x_{2n} \}\), \(\{ x_{2n-1} \}\), and \(\{ x_{3n} \}\) all converge. Show that \(\{ x_n \}\) is convergent.

### Facts about limits of sequences

*Note: 2–2.5 lectures, recursively defined sequences can safely be skipped*

In this section we go over some basic results about the limits of sequences. We start by looking at how sequences interact with inequalities.

#### Limits and inequalities

A basic lemma about limits and inequalities is the so-called squeeze lemma. It allows us to show convergence of sequences in difficult cases if we find two other simpler convergent sequences that “squeeze” the original sequence.

[squeeze:lemma] Let \(\{ a_n \}\), \(\{ b_n \}\), and \(\{ x_n \}\) be sequences such that \[a_n \leq x_n \leq b_n \quad \text{ for all $n \in {\mathbb{N}}$} .\] Suppose \(\{ a_n \}\) and \(\{ b_n \}\) converge and \[\lim_{n\to \infty} a_n = \lim_{n\to \infty} b_n .\] Then \(\{ x_n \}\) converges and \[\lim_{n\to \infty} x_n = \lim_{n\to \infty} a_n = \lim_{n\to \infty} b_n .\]

The intuitive idea of the proof is illustrated in . If \(x\) is the limit of \(a_n\) and \(b_n\), then if they are both within \(\nicefrac{\epsilon}{3}\) of \(x\), then the distance between \(a_n\) and \(b_n\) is at most \(\nicefrac{2\epsilon}{3}\). As \(x_n\) is between \(a_n\) and \(b_n\) it is at most \(\nicefrac{2\epsilon}{3}\) from \(a_n\). Since \(a_n\) is at most \(\nicefrac{\epsilon}{3}\) away from \(x\), then \(x_n\) must be at most \(\epsilon\) away from \(x\). Let us follow through on this intuition rigorously.

Let \(x := \lim\, a_n = \lim\, b_n\). Let \(\epsilon > 0\) be given.

Find an \(M_1\) such that for all \(n \geq M_1\) we have that \(\left\lvert {a_n-x} \right\rvert < \nicefrac{\epsilon}{3}\), and an \(M_2\) such that for all \(n \geq M_2\) we have \(\left\lvert {b_n-x} \right\rvert < \nicefrac{\epsilon}{3}\). Set \(M := \max \{M_1, M_2 \}\). Suppose \(n \geq M\). We compute \[\begin{split} \left\lvert {x_n - a_n} \right\rvert = x_n-a_n & \leq b_n-a_n \\ & = \left\lvert {b_n - x + x - a_n} \right\rvert \\ & \leq \left\lvert {b_n - x} \right\rvert + \left\lvert {x - a_n} \right\rvert \\ & < \frac{\epsilon}{3} + \frac{\epsilon}{3} = \frac{2\epsilon}{3} . \end{split}\] Armed with this information we estimate \[\begin{split} \left\lvert {x_n - x} \right\rvert &= \left\lvert {x_n - x + a_n - a_n} \right\rvert \\ &\leq \left\lvert {x_n - a_n} \right\rvert + \left\lvert {a_n - x} \right\rvert \\ & < \frac{2\epsilon}{3} + \frac{\epsilon}{3} = \epsilon . \end{split}\] And we are done.

One application of the is to compute limits of sequences using limits that are already known. For example, suppose we have the sequence \(\{ \frac{1}{n\sqrt{n}} \}\). Since \(\sqrt{n} \geq 1\) for all \(n \in {\mathbb{N}}\), we have \[0 \leq \frac{1}{n\sqrt{n}} \leq \frac{1}{n}\] for all \(n \in {\mathbb{N}}\). We already know \(\lim \nicefrac{1}{n} = 0\). Hence, using the constant sequence \(\{ 0 \}\) and the sequence \(\{ \nicefrac{1}{n} \}\) in the squeeze lemma, we conclude \[\lim_{n\to\infty} \frac{1}{n\sqrt{n}} = 0 .\]

Limits also preserve inequalities.

[limandineq:lemma] Let \(\{ x_n \}\) and \(\{ y_n \}\) be convergent sequences and \[x_n \leq y_n ,\] for all \(n \in {\mathbb{N}}\). Then \[\lim_{n\to\infty} x_n \leq \lim_{n\to\infty} y_n .\]

Let \(x := \lim\, x_n\) and \(y := \lim\, y_n\). Let \(\epsilon > 0\) be given. Find an \(M_1\) such that for all \(n \geq M_1\) we have \(\left\lvert {x_n-x} \right\rvert < \nicefrac{\epsilon}{2}\). Find an \(M_2\) such that for all \(n \geq M_2\) we have \(\left\lvert {y_n-y} \right\rvert < \nicefrac{\epsilon}{2}\). In particular, for some \(n \geq \max\{ M_1, M_2 \}\) we have \(x-x_n < \nicefrac{\epsilon}{2}\) and \(y_n-y < \nicefrac{\epsilon}{2}\). We add these inequalities to obtain \[y_n-x_n+x-y < \epsilon, \qquad \text{or} \qquad y_n-x_n < y-x+ \epsilon .\] Since \(x_n \leq y_n\) we have \(0 \leq y_n-x_n\) and hence \(0 < y-x+ \epsilon\). In other words \[x-y < \epsilon .\] Because \(\epsilon > 0\) was arbitrary we obtain \(x-y \leq 0\), as we have seen that a nonnegative number less than any positive \(\epsilon\) is zero. Therefore \(x \leq y\).

An easy corollary is proved using constant sequences in . The proof is left as an exercise.

[limandineq:cor]

- Let \(\{ x_n \}\) be a convergent sequence such that \(x_n \geq 0\), then \[\lim_{n\to\infty} x_n \geq 0.\]
- Let \(a,b \in {\mathbb{R}}\) and let \(\{ x_n \}\) be a convergent sequence such that \[a \leq x_n \leq b ,\] for all \(n \in {\mathbb{N}}\). Then \[a \leq \lim_{n\to\infty} x_n \leq b.\]

In and we cannot simply replace all the non-strict inequalities with strict inequalities. For example, let \(x_n := \nicefrac{-1}{n}\) and \(y_n := \nicefrac{1}{n}\). Then \(x_n < y_n\), \(x_n < 0\), and \(y_n > 0\) for all \(n\). However, these inequalities are not preserved by the limit operation as we have \(\lim\, x_n = \lim\, y_n = 0\). The moral of this example is that strict inequalities may become non-strict inequalities when limits are applied; if we know \(x_n < y_n\) for all \(n\), we may only conclude \[\lim_{n \to \infty} x_n \leq \lim_{n \to \infty} y_n .\] This issue is a common source of errors.

#### Continuity of algebraic operations

Limits interact nicely with algebraic operations.

[prop:contalg] Let \(\{ x_n \}\) and \(\{ y_n \}\) be convergent sequences.

- [prop:contalg:i] The sequence \(\{ z_n \}\), where \(z_n := x_n + y_n\), converges and \[\lim_{n \to \infty} (x_n + y_n) = \lim_{n \to \infty} z_n = \lim_{n \to \infty} x_n + \lim_{n \to \infty} y_n .\]
- [prop:contalg:ii] The sequence \(\{ z_n \}\), where \(z_n := x_n - y_n\), converges and \[\lim_{n \to \infty} (x_n - y_n) = \lim_{n \to \infty} z_n = \lim_{n \to \infty} x_n - \lim_{n \to \infty} y_n .\]
- [prop:contalg:iii] The sequence \(\{ z_n \}\), where \(z_n := x_n y_n\), converges and \[\lim_{n \to \infty} (x_n y_n) = \lim_{n \to \infty} z_n = \left( \lim_{n \to \infty} x_n \right) \left( \lim_{n \to \infty} y_n \right) .\]
- [prop:contalg:iv] If \(\lim\, y_n \not= 0\) and \(y_n \not= 0\) for all \(n \in {\mathbb{N}}\), then the sequence \(\{ z_n \}\), where \(z_n := \dfrac{x_n}{y_n}\), converges and \[\lim_{n \to \infty} \frac{x_n}{y_n} = \lim_{n \to \infty} z_n = %\frac{\lim_{n \to \infty} x_n}{\lim_{n \to \infty} y_n} . \frac{\lim\, x_n}{\lim\, y_n} .\]

Let us start with [prop:contalg:i]. Suppose \(\{ x_n \}\) and \(\{ y_n \}\) are convergent sequences and write \(z_n := x_n + y_n\). Let \(x := \lim\, x_n\), \(y := \lim\, y_n\), and \(z := x+y\).

Let \(\epsilon > 0\) be given. Find an \(M_1\) such that for all \(n \geq M_1\) we have \(\left\lvert {x_n - x} \right\rvert < \nicefrac{\epsilon}{2}\). Find an \(M_2\) such that for all \(n \geq M_2\) we have \(\left\lvert {y_n - y} \right\rvert < \nicefrac{\epsilon}{2}\). Take \(M := \max \{ M_1, M_2 \}\). For all \(n \geq M\) we have \[\begin{split} \left\lvert {z_n - z} \right\rvert &= \left\lvert {(x_n+y_n) - (x+y)} \right\rvert = \left\lvert {x_n-x + y_n-y} \right\rvert \\ & \leq \left\lvert {x_n-x} \right\rvert + \left\lvert {y_n-y} \right\rvert \\ & < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon. \end{split}\] Therefore [prop:contalg:i] is proved. Proof of [prop:contalg:ii] is almost identical and is left as an exercise.

Let us tackle [prop:contalg:iii]. Suppose again that \(\{ x_n \}\) and \(\{ y_n \}\) are convergent sequences and write \(z_n := x_n y_n\). Let \(x := \lim\, x_n\), \(y := \lim\, y_n\), and \(z := xy\).

Let \(\epsilon > 0\) be given. As \(\{ x_n \}\) is convergent, it is bounded. Therefore, find a \(B >0\) such that \(\left\lvert {x_n} \right\rvert \leq B\) for all \(n \in {\mathbb{N}}\). Find an \(M_1\) such that for all \(n \geq M_1\) we have \(\left\lvert {x_n - x} \right\rvert < \frac{\epsilon}{2(\left\lvert {y} \right\rvert+1)}\). Find an \(M_2\) such that for all \(n \geq M_2\) we have \(\left\lvert {y_n - y} \right\rvert < \frac{\epsilon}{2B}\). Take \(M := \max \{ M_1, M_2 \}\). For all \(n \geq M\) we have \[\begin{split} \left\lvert {z_n - z} \right\rvert &= \left\lvert {(x_ny_n) - (xy)} \right\rvert \\ & = \left\lvert {x_ny_n - (x+x_n-x_n)y} \right\rvert \\ & = \left\lvert {x_n(y_n -y) + (x_n - x)y} \right\rvert \\ & \leq \left\lvert {x_n(y_n -y)} \right\rvert + \left\lvert {(x_n - x)y} \right\rvert \\ & = \left\lvert {x_n} \right\rvert\left\lvert {y_n -y} \right\rvert + \left\lvert {x_n - x} \right\rvert\left\lvert {y} \right\rvert \\ & \leq B\left\lvert {y_n -y} \right\rvert + \left\lvert {x_n - x} \right\rvert\left\lvert {y} \right\rvert \\ & < B\frac{\epsilon}{2B} + \frac{\epsilon}{2(\left\lvert {y} \right\rvert+1)}\left\lvert {y} \right\rvert \\ & < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon . \end{split}\]

Finally let us tackle [prop:contalg:iv]. Instead of proving [prop:contalg:iv] directly, we prove the following simpler claim:

*Claim: If \(\{ y_n \}\) is a convergent sequence such that \(\lim\, y_n \not= 0\) and \(y_n \not= 0\) for all \(n \in {\mathbb{N}}\), then* \[\lim_{n\to\infty} \frac{1}{y_n} = \frac{1}{\lim\, y_n} .\]

Once the claim is proved, we take the sequence \(\{ \nicefrac{1}{y_n} \}\), multiply it by the sequence \(\{ x_n \}\) and apply item [prop:contalg:iii].

Proof of claim: Let \(\epsilon > 0\) be given. Let \(y := \lim\, y_n\). Find an \(M\) such that for all \(n \geq M\) we have \[\left\lvert {y_n - y} \right\rvert < \min \left\{ \left\lvert {y} \right\rvert^2\frac{\epsilon}{2}, \, \frac{\left\lvert {y} \right\rvert}{2} \right\} .\] Notice that we can make this claim as the right hand side is positive because \(\left\lvert {y} \right\rvert \not= 0\). Therefore for all \(n \geq M\) we have \(\left\lvert {y - y_n} \right\rvert < \frac{\left\lvert {y} \right\rvert}{2}\), and so \[\left\lvert {y} \right\rvert = \left\lvert {y - y_n + y_n } \right\rvert \leq \left\lvert {y - y_n} \right\rvert + \left\lvert { y_n } \right\rvert < \frac{\left\lvert {y} \right\rvert}{2} + \left\lvert {y_n} \right\rvert.\] Subtracting \(\nicefrac{\left\lvert {y} \right\rvert}{2}\) from both sides we obtain \(\nicefrac{\left\lvert {y} \right\rvert}{2} < \left\lvert {y_n} \right\rvert\), or in other words, \[\frac{1}{\left\lvert {y_n} \right\rvert} < \frac{2}{\left\lvert {y} \right\rvert} .\] Now we finish the proof of the claim: \[\begin{split} \left\lvert {\frac{1}{y_n} - \frac{1}{y}} \right\rvert &= \left\lvert {\frac{y - y_n}{y y_n}} \right\rvert \\ & = \frac{\left\lvert {y - y_n} \right\rvert}{\left\lvert {y} \right\rvert \left\lvert {y_n} \right\rvert} \\ & < \frac{\left\lvert {y - y_n} \right\rvert}{\left\lvert {y} \right\rvert} \, \frac{2}{\left\lvert {y} \right\rvert} \\ & < \frac{\left\lvert {y} \right\rvert^2 \frac{\epsilon}{2}}{\left\lvert {y} \right\rvert} \, \frac{2}{\left\lvert {y} \right\rvert} = \epsilon . \end{split}\] And we are done.

By plugging in constant sequences, we get several easy corollaries. If \(c \in {\mathbb{R}}\) and \(\{ x_n \}\) is a convergent sequence, then for example \[\lim_{n \to \infty} c x_n = c \left( \lim_{n \to \infty} x_n \right) \qquad \text{and} \qquad \lim_{n \to \infty} (c + x_n) = c + \lim_{n \to \infty} x_n .\] Similarly with constant subtraction and division.

As we can take limits past multiplication we can show (exercise) that \(\lim\, x_n^k = {(\lim\, x_n)}^k\) for all \(k \in {\mathbb{N}}\). That is, we can take limits past powers. Let us see if we can do the same with roots.

Let \(\{ x_n \}\) be a convergent sequence such that \(x_n \geq 0\). Then \[\lim_{n\to\infty} \sqrt{x_n} = \sqrt{ \lim_{n\to\infty} x_n } .\]

Of course to even make this statement, we need to apply to show that \(\lim\, x_n \geq 0\), so that we can take the square root without worry.

Let \(\{ x_n \}\) be a convergent sequence and let \(x := \lim\, x_n\).

First suppose \(x=0\). Let \(\epsilon > 0\) be given. Then there is an \(M\) such that for all \(n \geq M\) we have \(x_n = \left\lvert {x_n} \right\rvert < \epsilon^2\), or in other words \(\sqrt{x_n} < \epsilon\). Hence \[\left\lvert {\sqrt{x_n} - \sqrt{x}} \right\rvert = \sqrt{x_n} < \epsilon.\]

Now suppose \(x > 0\) (and hence \(\sqrt{x} > 0\)). \[\begin{split} \left\lvert {\sqrt{x_n}-\sqrt{x}} \right\rvert &= \left\lvert {\frac{x_n-x}{\sqrt{x_n}+\sqrt{x}}} \right\rvert \\ &= \frac{1}{\sqrt{x_n}+\sqrt{x}} \left\lvert {x_n-x} \right\rvert \\ & \leq \frac{1}{\sqrt{x}} \left\lvert {x_n-x} \right\rvert . \end{split}\] We leave the rest of the proof to the reader.

A similar proof works for the \(k\)th root. That is, we also obtain \(\lim\, x_n^{1/k} = {( \lim\, x_n )}^{1/k}\). We leave this to the reader as a challenging exercise.

We may also want to take the limit past the absolute value sign. The converse of this proposition is not true, see part b).

If \(\{ x_n \}\) is a convergent sequence, then \(\{ \left\lvert {x_n} \right\rvert \}\) is convergent and \[\lim_{n\to\infty} \left\lvert {x_n} \right\rvert = \left\lvert {\lim_{n\to\infty} x_n} \right\rvert .\]

We simply note the reverse triangle inequality \[\big\lvert \left\lvert {x_n} \right\rvert - \left\lvert {x} \right\rvert \big\rvert \leq \left\lvert {x_n-x} \right\rvert .\] Hence if \(\left\lvert {x_n -x} \right\rvert\) can be made arbitrarily small, so can \(\big\lvert \left\lvert {x_n} \right\rvert - \left\lvert {x} \right\rvert \big\rvert\). Details are left to the reader.

Let us see an example putting the above propositions together. Since we know that \(\lim \nicefrac{1}{n} = 0\), then \[\lim_{n\to \infty} \left\lvert {\sqrt{1 + \nicefrac{1}{n}} - \nicefrac{100}{n^2}} \right\rvert = \left\lvert {\sqrt{1 + (\lim \nicefrac{1}{n})} - 100 (\lim \nicefrac{1}{n})(\lim \nicefrac{1}{n})} \right\rvert = 1.\] That is, the limit on the left hand side exists because the right hand side exists. You really should read the above equality from right to left.

#### Recursively defined sequences

Now that we know we can interchange limits and algebraic operations, we can compute the limits of many sequences. One such class are recursively defined sequences, that is, sequences where the next number in the sequence computed using a formula from a fixed number of preceding elements in the sequence.

Let \(\{ x_n \}\) be defined by \(x_1 := 2\) and \[x_{n+1} := x_n - \frac{x_n^2-2}{2x_n} .\] We must first find out if this sequence is well defined; we must show we never divide by zero. Then we must find out if the sequence converges. Only then can we attempt to find the limit.

First let us prove \(x_n\) exists and \(x_n > 0\) for all \(n\) (so the sequence is well defined and bounded below). Let us show this by . We know that \(x_1 = 2 > 0\). For the induction step, suppose \(x_n > 0\). Then \[x_{n+1} = x_n - \frac{x_n^2-2}{2x_n} = \frac{2x_n^2 - x_n^2+2}{2x_n} = \frac{x_n^2+2}{2x_n} .\] If \(x_n > 0\), then \(x_n^2+2 > 0\) and hence \(x_{n+1} > 0\).

Next let us show that the sequence is monotone decreasing. If we show that \(x_n^2-2 \geq 0\) for all \(n\), then \(x_{n+1} \leq x_n\) for all \(n\). Obviously \(x_1^2-2 = 4-2 = 2 > 0\). For an arbitrary \(n\) we have \[x_{n+1}^2-2 = {\left( \frac{x_n^2+2}{2x_n} \right)}^2 - 2 = \frac{x_n^4+4x_n^2+4 - 8x_n^2}{4x_n^2} = \frac{x_n^4-4x_n^2+4}{4x_n^2} = \frac{{\left( x_n^2-2 \right)}^2}{4x_n^2} .\] Since any number squared is nonnegative, we have that \(x_{n+1}^2-2 \geq 0\) for all \(n\). Therefore, \(\{ x_n \}\) is monotone decreasing and bounded (\(x_n > 0\) for all \(n\)), and the limit exists. It remains to find the limit.

Let us write \[2x_nx_{n+1} = x_n^2+2 .\] Since \(\{ x_{n+1} \}\) is the 1-tail of \(\{ x_n \}\), it converges to the same limit. Let us define \(x := \lim\, x_n\). We take the limit of both sides to obtain \[2x^2 = x^2+2 ,\] or \(x^2 = 2\). As \(x_n > 0\) for all \(n\) we get \(x \geq 0\), and therefore \(x = \sqrt{2}\).

You may have seen the above sequence before. It is the *Newton’s method* ^{8} for finding the square root of 2. This method comes up very often in practice and converges very rapidly. Notice that we have used the fact that \(x_1^2 -2 >0\), although it was not strictly needed to show convergence by considering a tail of the sequence. In fact the sequence converges as long as \(x_1 \not= 0\), although with a negative \(x_1\) we would arrive at \(x=-\sqrt{2}\). By replacing the 2 in the numerator we obtain the square root of any positive number. These statements are left as an exercise.

You should, however, be careful. Before taking any limits, you must make sure the sequence converges. Let us see an example.

Suppose \(x_1 := 1\) and \(x_{n+1} := x_n^2+x_n\). If we blindly assumed that the limit exists (call it \(x\)), then we would get the equation \(x = x^2+x\), from which we might conclude \(x=0\). However, it is not hard to show that \(\{ x_n \}\) is unbounded and therefore does not converge.

The thing to notice in this example is that the method still works, but it depends on the initial value \(x_1\). If we set \(x_1 := 0\), then the sequence converges and the limit really is 0. An entire branch of mathematics, called dynamics, deals precisely with these issues.

#### Some convergence tests

It is not always necessary to go back to the definition of convergence to prove that a sequence is convergent. We first give a simple convergence test. The main idea is that \(\{ x_n \}\) converges to \(x\) if and only if \(\{ \left\lvert { x_n - x } \right\rvert \}\) converges to zero.

[convzero:prop] Let \(\{ x_n \}\) be a sequence. Suppose there is an \(x \in {\mathbb{R}}\) and a convergent sequence \(\{ a_n \}\) such that \[\lim_{n\to\infty} a_n = 0\] and \[\left\lvert {x_n - x} \right\rvert \leq a_n\] for all \(n\). Then \(\{ x_n \}\) converges and \(\lim\, x_n = x\).

Let \(\epsilon > 0\) be given. Note that \(a_n \geq 0\) for all \(n\). Find an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \(a_n = \left\lvert {a_n - 0} \right\rvert < \epsilon\). Then, for all \(n \geq M\) we have \[\left\lvert {x_n - x} \right\rvert \leq a_n < \epsilon . \qedhere\]

As the proposition shows, to study when a sequence has a limit is the same as studying when another sequence goes to zero. In general it may be hard to decide if a sequence converges, but for certain sequences there exist easy to apply tests that tell us if the sequence converges or not. Let us see one such test. First let us compute the limit of a very specific sequence.

Let \(c > 0\).

- If \(c < 1\), then \[\lim_{n\to\infty} c^n = 0.\]
- If \(c > 1\), then \(\{ c^n \}\) is unbounded.

First let us suppose \(c > 1\). We write \(c = 1+r\) for some \(r > 0\). By (or using the binomial theorem if you know it) we have Bernoulli’s inequality (see also ): \[c^n = {(1+r)}^n \geq 1+nr .\] By the of the real numbers, the sequence \(\{ 1+nr \}\) is unbounded (for any number \(B\), we find an \(n \in {\mathbb{N}}\) such that \(nr \geq B-1\)). Therefore \(c^n\) is unbounded.

Now let \(c < 1\). Write \(c = \frac{1}{1+r}\), where \(r > 0\). Then \[c^n = \frac{1}{{(1+r)}^n} \leq \frac{1}{1+nr} \leq \frac{1}{r} \frac{1}{n} .\] As \(\{ \frac{1}{n} \}\) converges to zero, so does \(\{ \frac{1}{r} \frac{1}{n} \}\). Hence, \(\{ c^n \}\) converges to zero.

If we look at the above proposition, we note that the ratio of the \((n+1)\)th term and the \(n\)th term is \(c\). We generalize this simple result to a larger class of sequences. The following lemma will come up again once we get to series.

[seq:ratiotest] Let \(\{ x_n \}\) be a sequence such that \(x_n \not= 0\) for all \(n\) and such that the limit \[L := \lim_{n\to\infty} \frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert}\] exists.

- If \(L < 1\), then \(\{ x_n \}\) converges and \(\lim\, x_n = 0\).
- If \(L > 1\), then \(\{ x_n \}\) is unbounded (hence diverges).

If \(L\) exists, but \(L=1\), the lemma says nothing. We cannot make any conclusion based on that information alone. For example, the sequence \(\{ \nicefrac{1}{n} \}\) converges to zero, but \(L=1\). The constant sequence \(\{ 1 \}\) converges to 1, not zero, and also \(L=1\). The sequence \(\{ {(-1)}^n \}\) does not converge at all, and \(L=1\). Finally the sequence \(\{ \ln n \}\) is unbounded, yet again \(L=1\).

Suppose \(L < 1\). As \(\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} \geq 0\), we have that \(L \geq 0\). Pick \(r\) such that \(L < r < 1\). We wish to compare the sequence to the sequence \(r^n\). The idea is that while the sequence is not going to be less than \(L\) eventually, it will eventually be less than \(r\), which is still less than 1. The intuitive idea of the proof is illustrated in .

As \(r-L > 0\), there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \[\left\lvert {\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} - L} \right\rvert < r-L .\] Therefore, \[\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} < r .\] For \(n > M\) (that is for \(n \geq M+1\)) we write \[\left\lvert {x_n} \right\rvert = \left\lvert {x_M} \right\rvert \frac{\left\lvert {x_{M+1}} \right\rvert}{\left\lvert {x_{M}} \right\rvert} \frac{\left\lvert {x_{M+2}} \right\rvert}{\left\lvert {x_{M+1}} \right\rvert} \cdots \frac{\left\lvert {x_{n}} \right\rvert}{\left\lvert {x_{n-1}} \right\rvert} < \left\lvert {x_M} \right\rvert r r \cdots r = \left\lvert {x_M} \right\rvert r^{n-M} = (\left\lvert {x_M} \right\rvert r^{-M}) r^n .\] The sequence \(\{ r^n \}\) converges to zero and hence \(\left\lvert {x_M} \right\rvert r^{-M} r^n\) converges to zero. By , the \(M\)-tail of \(\{x_n\}\) converges to zero and therefore \(\{x_n\}\) converges to zero.

Now suppose \(L > 1\). Pick \(r\) such that \(1 < r < L\). As \(L-r > 0\), there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \[\left\lvert {\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} - L} \right\rvert < L-r .\] Therefore, \[\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} > r .\] Again for \(n > M\) we write \[\left\lvert {x_n} \right\rvert = \left\lvert {x_M} \right\rvert \frac{\left\lvert {x_{M+1}} \right\rvert}{\left\lvert {x_{M}} \right\rvert} \frac{\left\lvert {x_{M+2}} \right\rvert}{\left\lvert {x_{M+1}} \right\rvert} \cdots \frac{\left\lvert {x_{n}} \right\rvert}{\left\lvert {x_{n-1}} \right\rvert} > \left\lvert {x_M} \right\rvert r r \cdots r = \left\lvert {x_M} \right\rvert r^{n-M} = (\left\lvert {x_M} \right\rvert r^{-M}) r^n .\] The sequence \(\{ r^n \}\) is unbounded (since \(r > 1\)), and therefore \(\{x_n\}\) cannot be bounded (if \(\left\lvert {x_n} \right\rvert \leq B\) for all \(n\), then \(r^n < \frac{B}{\left\lvert {x_M} \right\rvert} r^{M}\) for all \(n\), which is impossible). Consequently, \(\{ x_n \}\) cannot converge.

A simple application of the above lemma is to prove that \[\lim_{n\to\infty} \frac{2^n}{n!} = 0 .\]

Proof: We find that \[\frac{2^{n+1} / (n+1)!}{2^n/n!} = \frac{2^{n+1}}{2^n}\frac{n!}{(n+1)!} = \frac{2}{n+1} .\] It is not hard to see that \(\{ \frac{2}{n+1} \}\) converges to zero. The conclusion follows by the lemma.

## Exercises

Prove . Hint: Use constant sequences and .

Prove part [prop:contalg:ii] of .

Prove that if \(\{ x_n \}\) is a convergent sequence, \(k \in {\mathbb{N}}\), then \[\lim_{n\to\infty} x_n^k = {\left( \lim_{n\to\infty} x_n \right)}^k .\] Hint: Use .

Suppose \(x_1 := \frac{1}{2}\) and \(x_{n+1} := x_n^2\). Show that \(\{ x_n \}\) converges and find \(\lim\, x_n\). Hint: You cannot divide by zero!

Let \(x_n := \frac{n-\cos(n)}{n}\). Use the to show that \(\{ x_n \}\) converges and find the limit.

Let \(x_n := \frac{1}{n^2}\) and \(y_n := \frac{1}{n}\). Define \(z_n := \frac{x_n}{y_n}\) and \(w_n := \frac{y_n}{x_n}\). Do \(\{ z_n \}\) and \(\{ w_n \}\) converge? What are the limits? Can you apply ? Why or why not?

True or false, prove or find a counterexample. If \(\{ x_n \}\) is a sequence such that \(\{ x_n^2 \}\) converges, then \(\{ x_n \}\) converges.

Show that \[\lim_{n\to\infty} \frac{n^2}{2^n} = 0 .\]

Suppose \(\{ x_n \}\) is a sequence and suppose for some \(x \in {\mathbb{R}}\), the limit \[L := \lim_{n \to \infty} \frac{\left\lvert {x_{n+1}-x} \right\rvert}{\left\lvert {x_n-x} \right\rvert}\] exists and \(L < 1\). Show that \(\{ x_n \}\) converges to \(x\).

Let \(\{ x_n \}\) be a convergent sequence such that \(x_n \geq 0\) and \(k \in {\mathbb{N}}\). Then \[\lim_{n\to\infty} x_n^{1/k} = {\left( \lim_{n\to\infty} x_n \right)}^{1/k} .\] Hint: Find an expression \(q\) such that \(\frac{x_n^{1/k}-x^{1/k}}{x_n-x} = \frac{1}{q}\).

Let \(r > 0\). Show that starting with any \(x_1 \not= 0\), the sequence defined by \[x_{n+1} := x_n - \frac{x_n^2-r}{2x_n}\] converges to \(\sqrt{r}\) if \(x_1 > 0\) and \(-\sqrt{r}\) if \(x_1 < 0\).

a) Suppose \(\{ a_n \}\) is a bounded sequence and \(\{ b_n \}\) is a sequence converging to 0. Show that \(\{ a_n b_n \}\) converges to 0.

b) Find an example where \(\{ a_n \}\) is unbounded, \(\{ b_n \}\) converges to 0, and \(\{ a_n b_n \}\) is not convergent.

c) Find an example where \(\{ a_n \}\) is bounded, \(\{ b_n \}\) converges to some \(x \not= 0\), and \(\{ a_n b_n \}\) is not convergent.

### Limit superior, limit inferior, and Bolzano-Weierstrass

*Note: 1–2 lectures, alternative proof of BW optional*

In this section we study bounded sequences and their subsequences. In particular we define the so-called limit superior and limit inferior of a bounded sequence and talk about limits of subsequences. Furthermore, we prove the Bolzano-Weierstrass theorem ^{9}, which is an indispensable tool in analysis.

We have seen that every convergent sequence is bounded, although there exist many bounded divergent sequences. For example, the sequence \(\{ {(-1)}^n \}\) is bounded, but it is divergent. All is not lost however and we can still compute certain limits with a bounded divergent sequence.

#### Upper and lower limits

There are ways of creating monotone sequences out of any sequence, and in this fashion we get the so-called *limit superior* and *limit inferior*. These limits always exist for bounded sequences.

If a sequence \(\{ x_n \}\) is bounded, then the set \(\{ x_k : k \in {\mathbb{N}}\}\) is bounded. Then for every \(n\) the set \(\{ x_k : k \geq n \}\) is also bounded (as it is a subset).

[liminflimsup:def] Let \(\{ x_n \}\) be a bounded sequence. Let \(a_n := \sup \{ x_k : k \geq n \}\) and \(b_n := \inf \{ x_k : k \geq n \}\). Define \[\begin{aligned} \limsup_{n \to \infty} \, x_n & := \lim_{n \to \infty} a_n , \\ \liminf_{n \to \infty} \, x_n & := \lim_{n \to \infty} b_n .\end{aligned}\]

For a bounded sequence, liminf and limsup always exist (see below). It is possible to define liminf and limsup for unbounded sequences if we allow \(\infty\) and \(-\infty\). It is not hard to generalize the following results to include unbounded sequences, however, we first restrict our attention to bounded ones.

Let \(\{ x_n \}\) be a bounded sequence. Let \(a_n\) and \(b_n\) be as in the definition above.

- The sequence \(\{ a_n \}\) is bounded monotone decreasing and \(\{ b_n \}\) is bounded monotone increasing. In particular, \(\liminf x_n\) and \(\limsup x_n\) exist.
- \(\displaystyle \limsup_{n \to \infty} \, x_n = \inf \{ a_n : n \in {\mathbb{N}}\}\) and \(\displaystyle \liminf_{n \to \infty} \, x_n = \sup \{ b_n : n \in {\mathbb{N}}\}\).
- \(\displaystyle \liminf_{n \to \infty} \, x_n \leq \limsup_{n \to \infty} \, x_n\).

Let us see why \(\{ a_n \}\) is a decreasing sequence. As \(a_n\) is the least upper bound for \(\{ x_k : k \geq n \}\), it is also an upper bound for the subset \(\{ x_k : k \geq (n+1) \}\). Therefore \(a_{n+1}\), the least upper bound for \(\{ x_k : k \geq (n+1) \}\), has to be less than or equal to \(a_n\), that is, \(a_n \geq a_{n+1}\). Similarly (an exercise), \(b_n\) is an increasing sequence. It is left as an exercise to show that if \(x_n\) is bounded, then \(a_n\) and \(b_n\) must be bounded.

The second item in the proposition follows as the sequences \(\{ a_n \}\) and \(\{ b_n \}\) are monotone.

For the third item, we note that \(b_n \leq a_n\), as the \(\inf\) of a set is less than or equal to its \(\sup\). We know that \(\{ a_n \}\) and \(\{ b_n \}\) converge to the limsup and the liminf (respectively). We apply to obtain \[\lim_{n\to \infty} b_n \leq \lim_{n\to \infty} a_n. \qedhere\]

Let \(\{ x_n \}\) be defined by \[x_n := \begin{cases} \frac{n+1}{n} & \text{ if $n$ is odd,} \\ 0 & \text{ if $n$ is even.} \end{cases}\] Let us compute the \(\liminf\) and \(\limsup\) of this sequence. First the limit inferior: \[\liminf_{n\to\infty} \, x_n = \lim_{n\to\infty} \left( \inf \{ x_k : k \geq n \} \right) = \lim_{n\to\infty} 0 = 0 .\] For the limit superior we write \[\limsup_{n\to\infty} \, x_n = \lim_{n\to\infty} \left( \sup \{ x_k : k \geq n \} \right) .\] It is not hard to see that \[\sup \{ x_k : k \geq n \} = \begin{cases} \frac{n+1}{n} & \text{ if $n$ is odd,} \\ \frac{n+2}{n+1} & \text{ if $n$ is even.} \end{cases}\] We leave it to the reader to show that the limit is 1. That is, \[\limsup_{n\to\infty} \, x_n = 1 .\] Do note that the sequence \(\{ x_n \}\) is not a convergent sequence.

We associate with \(\limsup\) and \(\liminf\) certain subsequences.

[subseqlimsupinf:thm] If \(\{ x_n \}\) is a bounded sequence, then there exists a subsequence \(\{ x_{n_k} \}\) such that \[\lim_{k\to \infty} x_{n_k} = \limsup_{n \to \infty} \, x_n .\] Similarly, there exists a (perhaps different) subsequence \(\{ x_{m_k} \}\) such that \[\lim_{k\to \infty} x_{m_k} = \liminf_{n \to \infty} \, x_n .\]

Define \(a_n := \sup \{ x_k : k \geq n \}\). Write \(x := \limsup \, x_n = \lim\, a_n\). Define the subsequence as follows. Pick \(n_1 := 1\) and work inductively. Suppose we have defined the subsequence until \(n_k\) for some \(k\). Now pick some \(m > n_k\) such that \[a_{(n_k+1)} - x_m < \frac{1}{k+1} .\] We can do this as \(a_{(n_k+1)}\) is a supremum of the set \(\{ x_n : n \geq n_k + 1 \}\) and hence there are elements of the sequence arbitrarily close (or even possibly equal) to the supremum. Set \(n_{k+1} := m\). The subsequence \(\{ x_{n_k} \}\) is defined. Next we need to prove that it converges and has the right limit.

Note that \(a_{(n_{k-1}+1)} \geq a_{n_k}\) (why?) and that \(a_{n_{k}} \geq x_{n_k}\). Therefore, for every \(k > 1\) we have \[\begin{split} \left\lvert {a_{n_k} - x_{n_k}} \right\rvert & = a_{n_k} - x_{n_k} \\ & \leq a_{(n_{k-1}+1)} - x_{n_k} \\ & < \frac{1}{k} . \end{split}\]

Let us show that \(\{ x_{n_k} \}\) converges to \(x\). Note that the subsequence need not be monotone. Let \(\epsilon > 0\) be given. As \(\{ a_n \}\) converges to \(x\), then the subsequence \(\{ a_{n_k} \}\) converges to \(x\). Thus there exists an \(M_1 \in {\mathbb{N}}\) such that for all \(k \geq M_1\) we have \[\left\lvert {a_{n_k} - x} \right\rvert < \frac{\epsilon}{2} .\] Find an \(M_2 \in {\mathbb{N}}\) such that \[\frac{1}{M_2} \leq \frac{\epsilon}{2}.\] Take \(M := \max \{M_1 , M_2 \}\) and compute. For all \(k \geq M\) we have \[\begin{split} \left\lvert {x- x_{n_k}} \right\rvert & = \left\lvert {a_{n_k} - x_{n_k} + x - a_{n_k}} \right\rvert \\ & \leq \left\lvert {a_{n_k} - x_{n_k}} \right\rvert + \left\lvert {x - a_{n_k}} \right\rvert \\ & < \frac{1}{k} + \frac{\epsilon}{2} \\ & \leq \frac{1}{M_2} + \frac{\epsilon}{2} \leq \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon . \end{split}\]

We leave the statement for \(\liminf\) as an exercise.

#### Using limit inferior and limit superior

The advantage of \(\liminf\) and \(\limsup\) is that we can always write them down for any (bounded) sequence. If we could somehow compute them, we could also compute the limit of the sequence if it exists, or show that the sequence diverges. Working with \(\liminf\) and \(\limsup\) is a little bit like working with limits, although there are subtle differences.

[liminfsupconv:thm] Let \(\{ x_n \}\) be a bounded sequence. Then \(\{ x_n \}\) converges if and only if \[\liminf_{n\to \infty} \, x_n = \limsup_{n\to \infty} \, x_n.\] Furthermore, if \(\{ x_n \}\) converges, then \[\lim_{n\to \infty} x_n = \liminf_{n\to \infty} \, x_n = \limsup_{n\to \infty} \, x_n.\]

Define \(a_n\) and \(b_n\) as in . Note that \[b_n \leq x_n \leq a_n .\] If \(\liminf \, x_n = \limsup \, x_n\), then we know that \(\{ a_n \}\) and \(\{ b_n \}\) have limits and that these two limits are the same. By the squeeze lemma (), \(\{ x_n \}\) converges and \[\lim_{n\to \infty} b_n = \lim_{n\to \infty} x_n = \lim_{n\to \infty} a_n .\]

Now suppose \(\{ x_n \}\) converges to \(x\). We know by that there exists a subsequence \(\{ x_{n_k} \}\) that converges to \(\limsup \, x_n\). As \(\{ x_n \}\) converges to \(x\), every subsequence converges to \(x\) and therefore \(\limsup \, x_n = \lim\, x_{n_k} = x\). Similarly \(\liminf \, x_n = x\).

Limit superior and limit inferior behave nicely with subsequences.

[prop:subseqslimsupinf] Suppose \(\{ x_n \}\) is a bounded sequence and \(\{ x_{n_k} \}\) is a subsequence. Then \[\liminf_{n\to\infty} \, x_n \leq \liminf_{k\to\infty} \, x_{n_k} \leq \limsup_{k\to\infty} \, x_{n_k} \leq \limsup_{n\to\infty} \, x_n .\]

The middle inequality has been proved already. We will prove the third inequality, and leave the first inequality as an exercise.

We want to prove that \(\limsup \, x_{n_k} \leq \limsup \, x_n\). Define \(a_j := \sup \{ x_k : k \geq j \}\) as usual. Also define \(c_j := \sup \{ x_{n_k} : k \geq j \}\). It is not true that \(c_j\) is necessarily a subsequence of \(a_j\). However, as \(n_k \geq k\) for all \(k\), we have that \(\{ x_{n_k} : k \geq j \} \subset \{ x_k : k \geq j \}\). A supremum of a subset is less than or equal to the supremum of the set and therefore \[c_j \leq a_j .\] We apply to conclude \[\lim_{j\to\infty} c_j \leq \lim_{j\to\infty} a_j ,\] which is the desired conclusion.

Limit superior and limit inferior are the largest and smallest subsequential limits. If the subsequence in the previous proposition is convergent, then we have that \(\liminf \, x_{n_k} = \lim\, x_{n_k} = \limsup \, x_{n_k}\). Therefore, \[\liminf_{n\to\infty} \, x_n \leq \lim_{k\to\infty} x_{n_k} \leq \limsup_{n\to\infty} \, x_n .\]

Similarly we get the following useful test for convergence of a bounded sequence. We leave the proof as an exercise.

[seqconvsubseqconv:thm] A bounded sequence \(\{ x_n \}\) is convergent and converges to \(x\) if and only if every convergent subsequence \(\{ x_{n_k} \}\) converges to \(x\).

#### Bolzano-Weierstrass theorem

While it is not true that a bounded sequence is convergent, the Bolzano-Weierstrass theorem tells us that we can at least find a convergent subsequence. The version of Bolzano-Weierstrass that we present in this section is the Bolzano-Weierstrass for sequences.

[thm:bwseq] Suppose a sequence \(\{ x_n \}\) of real numbers is bounded. Then there exists a convergent subsequence \(\{ x_{n_i} \}\).

We use . It says that there exists a subsequence whose limit is \(\limsup \, x_n\).

The reader might complain right now that is strictly stronger than the Bolzano-Weierstrass theorem as presented above. That is true. However, only applies to the real line, but Bolzano-Weierstrass applies in more general contexts (that is, in \({\mathbb{R}}^n\)) with pretty much the exact same statement.

As the theorem is so important to analysis, we present an explicit proof. The following proof generalizes more easily to different contexts.

As the sequence is bounded, then there exist two numbers \(a_1 < b_1\) such that \(a_1 \leq x_n \leq b_1\) for all \(n \in {\mathbb{N}}\).

We will define a subsequence \(\{ x_{n_i} \}\) and two sequences \(\{ a_i \}\) and \(\{ b_i \}\), such that \(\{ a_i \}\) is monotone increasing, \(\{ b_i \}\) is monotone decreasing, \(a_i \leq x_{n_i} \leq b_i\) and such that \(\lim\, a_i = \lim\, b_i\). That \(x_{n_i}\) converges follows by the .

We define the sequences inductively. We will always have that \(a_i < b_i\), and that \(x_n \in [a_i,b_i]\) for infinitely many \(n \in {\mathbb{N}}\). We have already defined \(a_1\) and \(b_1\). We take \(n_1 := 1\), that is \(x_{n_1} = x_1\).

Now suppose that up to some \(k \in {\mathbb{N}}\) we have defined the subsequence \(x_{n_1}, x_{n_2}, \ldots, x_{n_k}\), and the sequences \(a_1,a_2,\ldots,a_k\) and \(b_1,b_2,\ldots,b_k\). Let \(y := \frac{a_k+b_k}{2}\). Clearly \(a_k < y < b_k\). If there exist infinitely many \(j \in {\mathbb{N}}\) such that \(x_j \in [a_k,y]\), then set \(a_{k+1} := a_k\), \(b_{k+1} := y\), and pick \(n_{k+1} > n_{k}\) such that \(x_{n_{k+1}} \in [a_k,y]\). If there are not infinitely many \(j\) such that \(x_j \in [a_k,y]\), then it must be true that there are infinitely many \(j \in {\mathbb{N}}\) such that \(x_j \in [y,b_k]\). In this case pick \(a_{k+1} := y\), \(b_{k+1} := b_k\), and pick \(n_{k+1} > n_{k}\) such that \(x_{n_{k+1}} \in [y,b_k]\).

Now we have the sequences defined. What is left to prove is that \(\lim\, a_i = \lim\, b_i\). Obviously the limits exist as the sequences are monotone. From the construction, it is obvious that \(b_i - a_i\) is cut in half in each step. Therefore \(b_{i+1} - a_{i+1} = \frac{b_i-a_i}{2}\). By , we obtain that \[b_i - a_i = \frac{b_1-a_1}{2^{i-1}} .\]

Let \(x := \lim\, a_i\). As \(\{ a_i \}\) is monotone we have that \[x = \sup \{ a_i : i \in {\mathbb{N}}\}\] Now let \(y := \lim\, b_i = \inf \{ b_i : i \in {\mathbb{N}}\}\). Obviously \(y \leq x\) as \(a_i < b_i\) for all \(i\). As the sequences are monotone, then for any \(i\) we have (why?) \[y-x \leq b_i-a_i = \frac{b_1-a_1}{2^{i-1}} .\] As \(\frac{b_1-a_1}{2^{i-1}}\) is arbitrarily small and \(y-x \geq 0\), we have that \(y-x = 0\). We finish by the .

Yet another proof of the Bolzano-Weierstrass theorem is to show the following claim, which is left as a challenging exercise. *Claim: Every sequence has a monotone subsequence*.

#### Infinite limits

If we allow \(\liminf\) and \(\limsup\) to take on the values \(\infty\) and \(-\infty\), we can apply \(\liminf\) and \(\limsup\) to all sequences, not just bounded ones. For any sequence, we write \[\limsup \, x_n := \inf \{ a_n : n \in {\mathbb{N}}\}, \qquad \text{and} \qquad \liminf \, x_n := \sup \{ b_n : n \in {\mathbb{N}}\},\] where \(a_n := \sup \{ x_k : k \geq n \}\) and \(b_n := \inf \{ x_k : k \geq n \}\) as before.

We also often define infinite limits for certain divergent sequences.

We say \(\{ x_n \}\) *diverges to infinity* ^{10} if for every \(M \in {\mathbb{R}}\), there exists an \(N \in {\mathbb{N}}\) such that for all \(n \geq N\) we have \(x_n > M\). In this case we write \(\lim \, x_n := \infty\). Similarly if for every \(M \in {\mathbb{R}}\) there exists an \(N \in {\mathbb{N}}\) such that for all \(n \geq N\) we have \(x_n < M\), we say \(\{ x_n \}\) *diverges to minus infinity* and we write \(\lim \, x_n := -\infty\).

This definition behaves as expected with \(\limsup\) and \(\liminf\), see exercises [exercise:infseqlimex] and [exercise:infseqlimlims].

If \(x_n := 0\) for odd \(n\) and \(x_n := n\) for even \(n\) then \[\lim_{n\to \infty} n = \infty, \qquad \lim_{n\to \infty} x_n \quad \text{does not exist}, \qquad \limsup_{n\to \infty} x_n = \infty.\]

## Exercises

Suppose \(\{ x_n \}\) is a bounded sequence. Define \(a_n\) and \(b_n\) as in . Show that \(\{ a_n \}\) and \(\{ b_n \}\) are bounded.

Suppose \(\{ x_n \}\) is a bounded sequence. Define \(b_n\) as in . Show that \(\{ b_n \}\) is an increasing sequence.

Finish the proof of . That is, suppose \(\{ x_n \}\) is a bounded sequence and \(\{ x_{n_k} \}\) is a subsequence. Prove \(\displaystyle \liminf_{n\to\infty}\, x_n \leq \liminf_{k\to\infty}\, x_{n_k}\).

Prove .

- Let \(x_n := \dfrac{{(-1)}^n}{n}\), find \(\limsup \, x_n\) and \(\liminf \, x_n\).
- Let \(x_n := \dfrac{(n-1){(-1)}^n}{n}\), find \(\limsup \, x_n\) and \(\liminf \, x_n\).

Let \(\{ x_n \}\) and \(\{ y_n \}\) be bounded sequences such that \(x_n \leq y_n\) for all \(n\). Then show that \[\limsup_{n\to\infty} \, x_n \leq \limsup_{n\to\infty} \, y_n\] and \[\liminf_{n\to\infty} \, x_n \leq \liminf_{n\to\infty} \, y_n .\]

Let \(\{ x_n \}\) and \(\{ y_n \}\) be bounded sequences.

- Show that \(\{ x_n + y_n \}\) is bounded.
- Show that \[(\liminf_{n\to \infty}\, x_n) + (\liminf_{n\to \infty}\, y_n) \leq \liminf_{n\to \infty}\, (x_n+y_n) .\] Hint: Find a subsequence \(\{ x_{n_i}+y_{n_i} \}\) of \(\{ x_n + y_n \}\) that converges. Then find a subsequence \(\{ x_{n_{m_i}} \}\) of \(\{ x_{n_i} \}\) that converges. Then apply what you know about limits.
- Find an explicit \(\{ x_n \}\) and \(\{ y_n \}\) such that \[(\liminf_{n\to \infty}\, x_n) + (\liminf_{n\to \infty}\, y_n) < \liminf_{n\to \infty}\, (x_n+y_n) .\] Hint: Look for examples that do not have a limit.

Let \(\{ x_n \}\) and \(\{ y_n \}\) be bounded sequences (from the previous exercise we know that \(\{ x_n + y_n \}\) is bounded).

- Show that \[(\limsup_{n\to \infty}\, x_n) + (\limsup_{n\to \infty}\, y_n) \geq \limsup_{n\to \infty}\, (x_n+y_n) .\] Hint: See previous exercise.
- Find an explicit \(\{ x_n \}\) and \(\{ y_n \}\) such that \[(\limsup_{n\to \infty}\, x_n) + (\limsup_{n\to \infty}\, y_n) > \limsup_{n\to \infty}\, (x_n+y_n) .\] Hint: See previous exercise.

If \(S \subset {\mathbb{R}}\) is a set, then \(x \in {\mathbb{R}}\) is a *cluster point* if for every \(\epsilon > 0\), the set \((x-\epsilon,x+\epsilon) \cap S \setminus \{ x \}\) is not empty. That is, if there are points of \(S\) arbitrarily close to \(x\). For example, \(S := \{ \nicefrac{1}{n} : n \in {\mathbb{N}}\}\) has a unique (only one) cluster point \(0\), but \(0 \notin S\). Prove the following version of the Bolzano-Weierstrass theorem:

* Theorem. Let \(S \subset {\mathbb{R}}\) be a bounded infinite set, then there exists at least one cluster point of \(S\)*.

Hint: If \(S\) is infinite, then \(S\) contains a countably infinite subset. That is, there is a sequence \(\{ x_n \}\) of distinct numbers in \(S\).

a) Prove that any sequence contains a monotone subsequence. Hint: Call \(n \in {\mathbb{N}}\) a *peak* if \(a_m \leq a_n\) for all \(m \geq n\). There are two possibilities: either the sequence has at most finitely many peaks, or it has infinitely many peaks.

b) Conclude the Bolzano-Weierstrass theorem.

Let us prove a stronger version of . Suppose \(\{ x_n \}\) is a sequence such that every subsequence \(\{ x_{n_i} \}\) has a subsequence \(\{ x_{n_{m_i}} \}\) that converges to \(x\). a) First show that \(\{ x_n \}\) is bounded. b) Now show that \(\{ x_n \}\) converges to \(x\).

Let \(\{x_n\}\) be a bounded sequence.

a) Prove that there exists an \(s\) such that for any \(r > s\) there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \(x_n < r\).

b) If \(s\) is a number as in a), then prove \(\limsup \, x_n \leq s\).

c) Show that if \(S\) is the set of all \(s\) as in a), then \(\limsup \, x_n = \inf \, S\).

[exercise:infseqlimex] Suppose \(\{ x_n \}\) is such that \(\liminf \, x_n = -\infty\), \(\limsup \, x_n = \infty\). a) Show that \(\{ x_n \}\) is not convergent, and also that neither \(\lim \, x_n = \infty\) nor \(\lim \, x_n = -\infty\) is true. b) Find an example of such a sequence.

[exercise:infseqlimlims] Given a sequence \(\{ x_n \}\). a) Show that \(\lim \, x_n = \infty\) if and only if \(\liminf \, x_n = \infty\). b) Then show that \(\lim \, x_n = - \infty\) if and only if \(\limsup \, x_n = -\infty\). c) If \(\{ x_n \}\) is monotone increasing, show that either \(\lim \, x_n\) exists and is finite or \(\lim \, x_n = \infty\).

### Cauchy sequences

*Note: 0.5–1 lecture*

Often we wish to describe a certain number by a sequence that converges to it. In this case, it is impossible to use the number itself in the proof that the sequence converges. It would be nice if we could check for convergence without knowing the limit.

A sequence \(\{ x_n \}\) is a *Cauchy sequence* ^{11} if for every \(\epsilon > 0\) there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) and all \(k \geq M\) we have \[\left\lvert {x_n - x_k} \right\rvert < \epsilon .\]

Intuitively what it means is that the terms of the sequence are eventually arbitrarily close to each other. We would expect such a sequence to be convergent. It turns out that is true because \({\mathbb{R}}\) has the . First, let us look at some examples.

The sequence \(\{ \nicefrac{1}{n} \}\) is a Cauchy sequence.

Proof: Given \(\epsilon > 0\), find \(M\) such that \(M > \nicefrac{2}{\epsilon}\). Then for \(n,k \geq M\) we have that \(\nicefrac{1}{n} < \nicefrac{\epsilon}{2}\) and \(\nicefrac{1}{k} < \nicefrac{\epsilon}{2}\). Therefore for \(n, k \geq M\) we have \[\left\lvert {\frac{1}{n} - \frac{1}{k}} \right\rvert \leq \left\lvert {\frac{1}{n}} \right\rvert + \left\lvert {\frac{1}{k}} \right\rvert < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon.\]

The sequence \(\{ \frac{n+1}{n} \}\) is a Cauchy sequence.

Proof: Given \(\epsilon > 0\), find \(M\) such that \(M > \nicefrac{2}{\epsilon}\). Then for \(n,k \geq M\) we have that \(\nicefrac{1}{n} < \nicefrac{\epsilon}{2}\) and \(\nicefrac{1}{k} < \nicefrac{\epsilon}{2}\). Therefore for \(n, k \geq M\) we have \[\begin{split} \left\lvert {\frac{n+1}{n} - \frac{k+1}{k}} \right\rvert & = \left\lvert {\frac{k(n+1)-n(k+1)}{nk}} \right\rvert \\ & = \left\lvert {\frac{kn+k-nk-n}{nk}} \right\rvert \\ & = \left\lvert {\frac{k-n}{nk}} \right\rvert \\ & \leq \left\lvert {\frac{k}{nk}} \right\rvert + \left\lvert {\frac{-n}{nk}} \right\rvert \\ & = \frac{1}{n} + \frac{1}{k} < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon . \end{split}\]

A Cauchy sequence is bounded.

Suppose \(\{ x_n \}\) is Cauchy. Pick \(M\) such that for all \(n,k \geq M\) we have \(\left\lvert {x_n-x_k} \right\rvert < 1\). In particular, we have that for all \(n \geq M\) \[\left\lvert {x_n - x_M} \right\rvert < 1 .\] Or by the reverse triangle inequality, \(\left\lvert {x_n} \right\rvert - \left\lvert {x_M} \right\rvert \leq \left\lvert {x_n - x_M} \right\rvert < 1\). Hence for \(n \geq M\) we have \[\left\lvert {x_n} \right\rvert < 1 + \left\lvert {x_M} \right\rvert.\] Let \[B := \max \{ \left\lvert {x_1} \right\rvert, \left\lvert {x_2} \right\rvert, \ldots, \left\lvert {x_{M-1}} \right\rvert, 1+ \left\lvert {x_M} \right\rvert \} .\] Then \(\left\lvert {x_n} \right\rvert \leq B\) for all \(n \in {\mathbb{N}}\).

A sequence of real numbers is Cauchy if and only if it converges.

Let \(\epsilon > 0\) be given and suppose \(\{ x_n \}\) converges to \(x\). Then there exists an \(M\) such that for \(n \geq M\) we have \[\left\lvert {x_n - x} \right\rvert < \frac{\epsilon}{2} .\] Hence for \(n \geq M\) and \(k \geq M\) we have \[\left\lvert {x_n - x_k} \right\rvert = \left\lvert {x_n - x + x - x_k} \right\rvert \leq \left\lvert {x_n-x} \right\rvert + \left\lvert {x-x_k} \right\rvert < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon .\]

Alright, that direction was easy. Now suppose \(\{ x_n \}\) is Cauchy. We have shown that \(\{ x_n \}\) is bounded. If we show that \[\liminf_{n\to \infty} \, x_n = \limsup_{n\to\infty} \, x_n ,\] then \(\{ x_n \}\) must be convergent by . Assuming that liminf and limsup exist is where we use the .

Define \(a := \limsup \, x_n\) and \(b := \liminf \, x_n\). By , there exist subsequences \(\{ x_{n_i} \}\) and \(\{ x_{m_i} \}\), such that \[\lim_{i\to\infty} x_{n_i} = a \qquad \text{and} \qquad \lim_{i\to\infty} x_{m_i} = b.\] Given an \(\epsilon > 0\), there exists an \(M_1\) such that for all \(i \geq M_1\) we have \(\left\lvert {x_{n_i} - a} \right\rvert < \nicefrac{\epsilon}{3}\) and an \(M_2\) such that for all \(i \geq M_2\) we have \(\left\lvert {x_{m_i} - b} \right\rvert < \nicefrac{\epsilon}{3}\). There also exists an \(M_3\) such that for all \(n,k \geq M_3\) we have \(\left\lvert {x_n-x_k} \right\rvert < \nicefrac{\epsilon}{3}\). Let \(M := \max \{ M_1, M_2, M_3 \}\). Note that if \(i \geq M\), then \(n_i \geq M\) and \(m_i \geq M\). Hence \[\begin{split} \left\lvert {a-b} \right\rvert & = \left\lvert {a-x_{n_i}+x_{n_i} -x_{m_i}+x_{m_i} -b} \right\rvert \\ & \leq \left\lvert {a-x_{n_i}} \right\rvert + \left\lvert {x_{n_i} -x_{m_i}} \right\rvert + \left\lvert {x_{m_i} -b} \right\rvert \\ & < \frac{\epsilon}{3} + \frac{\epsilon}{3} + \frac{\epsilon}{3} = \epsilon . \end{split}\] As \(\left\lvert {a-b} \right\rvert < \epsilon\) for all \(\epsilon > 0\), then \(a=b\) and the sequence converges.

The statement of this proposition is sometimes used to define the completeness property of the real numbers. We say a set is *Cauchy-complete* (or sometimes just *complete*) if every Cauchy sequence converges. Above we proved that as \({\mathbb{R}}\) has the , then \({\mathbb{R}}\) is Cauchy-complete. We can “complete” \({\mathbb{Q}}\) by “throwing in” just enough points to make all Cauchy sequences converge (we omit the details). The resulting field has the least-upper-bound property. The advantage of using Cauchy sequences to define completeness is that this idea generalizes to more abstract settings.

It should be noted that the Cauchy criterion is stronger than just \(\left\lvert {x_{n+1}-x_n} \right\rvert\) (or \(\left\lvert {x_{n+j}-x_n} \right\rvert\) for a fixed \(j\)) going to zero as \(n\) goes to infinity. In fact, when we get to the partial sums of the harmonic series (see in the next section), we will have a sequence such that \(x_{n+1}-x_n = \nicefrac{1}{n}\), yet \(\{ x_n \}\) is divergent. In fact, for that sequence it is true that \(\lim_{n\to\infty} \left\lvert {x_{n+j}-x_n} \right\rvert = 0\) for any \(j \in {\mathbb{N}}\) (confer ). The key point in the definition of Cauchy is that \(n\) and \(k\) vary independently and can be arbitrarily far apart.

## Exercises

Prove that \(\{ \frac{n^2-1}{n^2} \}\) is Cauchy using directly the definition of Cauchy sequences.

Let \(\{ x_n \}\) be a sequence such that there exists a \(0 < C < 1\) such that \[\left\lvert {x_{n+1} - x_n} \right\rvert \leq C \left\lvert {x_{n}-x_{n-1}} \right\rvert .\] Prove that \(\{ x_n \}\) is Cauchy. Hint: You can freely use the formula (for \(C \not= 1\)) \[1+ C+ C^2 + \cdots + C^n = \frac{1-C^{n+1}}{1-C}.\]

Suppose \(F\) is an ordered field that contains the rational numbers \({\mathbb{Q}}\), such that \({\mathbb{Q}}\) is dense, that is: whenever \(x,y \in F\) are such that \(x < y\), then there exists a \(q \in {\mathbb{Q}}\) such that \(x < q < y\). Say a sequence \(\{ x_n \}_{n=1}^\infty\) of rational numbers is Cauchy if given any \(\epsilon \in {\mathbb{Q}}\) with \(\epsilon > 0\), there exists an \(M\) such that for all \(n,k \geq M\) we have \(\left\lvert {x_n-x_k} \right\rvert < \epsilon\). Suppose any Cauchy sequence of rational numbers has a limit in \(F\). Prove that \(F\) has the .

Let \(\{ x_n \}\) and \(\{ y_n \}\) be sequences such that \(\lim\, y_n =0\). Suppose that for all \(k \in {\mathbb{N}}\) and for all \(m \geq k\) we have \[\left\lvert {x_m-x_k} \right\rvert \leq y_k .\] Show that \(\{ x_n \}\) is Cauchy.

Suppose a Cauchy sequence \(\{ x_n \}\) is such that for every \(M \in {\mathbb{N}}\), there exists a \(k \geq M\) and an \(n \geq M\) such that \(x_k < 0\) and \(x_n > 0\). Using simply the definition of a Cauchy sequence and of a convergent sequence, show that the sequence converges to \(0\).

Suppose \(\left\lvert {x_n-x_k} \right\rvert \leq \nicefrac{n}{k^2}\) for all \(n\) and \(k\). Show that \(\{ x_n \}\) is Cauchy.

Suppose \(\{ x_n \}\) is a Cauchy sequence such that for infinitely many \(n\), \(x_n = c\). Using only the definition of Cauchy sequence prove that \(\lim\, x_n = c\).

True/False prove or find a counterexample: If \(\{ x_n \}\) is a Cauchy sequence then there exists an \(M\) such that for all \(n \geq M\) we have \(\left\lvert {x_{n+1}-x_n} \right\rvert \leq \left\lvert {x_{n}-x_{n-1}} \right\rvert\).

### Series

*Note: 2 lectures*

A fundamental object in mathematics is that of a series. In fact, when foundations of analysis were being developed, the motivation was to understand series. Understanding series is very important in applications of analysis. For example, solving differential equations often includes series, and differential equations are the basis for understanding almost all of modern science.

#### Definition

Given a sequence \(\{ x_n \}\), we write the formal object \[\sum_{n=1}^\infty x_n \qquad \text{or sometimes just} \qquad \sum x_n\] and call it a *series*. A series *converges*, if the sequence \(\{ s_k \}\) defined by \[s_k := \sum_{n=1}^k x_n = x_1 + x_2 + \cdots + x_k ,\] converges. The numbers \(s_k\) are called *partial sums*. If \(x := \lim\, s_k\), we write \[\sum_{n=1}^\infty x_n = x .\] In this case, we cheat a little and treat \(\sum_{n=1}^\infty x_n\) as a number.

On the other hand, if the sequence \(\{ s_k \}\) diverges, we say the series is *divergent*. In this case, \(\sum x_n\) is simply a formal object and not a number.

In other words, for a convergent series we have \[\sum_{n=1}^\infty x_n = \lim_{k\to\infty} \sum_{n=1}^k x_n .\] We should be careful to only use this equality if the limit on the right actually exists. That is, the right-hand side does not make sense (the limit does not exist) if the series does not converge.

Before going further, let us remark that it is sometimes convenient to start the series at an index different from 1. That is, for example we can write \[\sum_{n=0}^\infty r^n = \sum_{n=1}^\infty r^{n-1} .\] The left-hand side is more convenient to write. The idea is the same as the notation for the tail of a sequence.

It is common to write the series \(\sum x_n\) as \[x_1 + x_2 + x_3 + \cdots\] with the understanding that the ellipsis indicates a series and not a simple sum. We do not use this notation as it often leads to mistakes in proofs.

The series \[\sum_{n=1}^\infty \frac{1}{2^n}\] converges and the limit is 1. That is, \[\sum_{n=1}^\infty \frac{1}{2^n} = \lim_{k\to\infty} \sum_{n=1}^k \frac{1}{2^n} = 1 .\]

Proof: First we prove the following equality \[\left( \sum_{n=1}^k \frac{1}{2^n} \right) + \frac{1}{2^k} = 1 .\] The equality is easy to see when \(k=1\). The proof for general \(k\) follows by , which we leave to the reader. Let \(s_k\) be the partial sum. We write \[\left\lvert { 1 - s_k } \right\rvert = \left\lvert { 1 - \sum_{n=1}^k \frac{1}{2^n} } \right\rvert = \left\lvert {\frac{1}{2^k}} \right\rvert = \frac{1}{2^k} .\] The sequence \(\{ \frac{1}{2^k} \}\) and therefore \(\{ \left\lvert {1-s_k} \right\rvert \}\) converges to zero. So, \(\{ s_k \}\) converges to 1.

For \(-1 < r < 1\), the *geometric series* \[\sum_{n=0}^\infty r^n\] converges. In fact, \(\sum_{n=0}^\infty r^n = \frac{1}{1-r}\). The proof is left as an exercise to the reader. The proof consists of showing \[\sum_{n=0}^{k-1} r^n = \frac{1-r^k}{1-r} ,\] and then taking the limit as \(k\) goes to \(\infty\).

A fact we often use is the following analogue of looking at the tail of a sequence.

Let \(\sum x_n\) be a series. Let \(M \in {\mathbb{N}}\). Then \[\sum_{n=1}^\infty x_n \quad \text{converges if and only if} \quad \sum_{n=M}^\infty x_n \quad \text{converges.}\]

We look at partial sums of the two series (for \(k \geq M\)) \[\sum_{n=1}^{k} x_n = \left( \sum_{n=1}^{M-1} x_n \right) + \sum_{n=M}^{k} x_n .\] Note that \(\sum_{n=1}^{M-1} x_n\) is a fixed number. Now use to finish the proof.

#### Cauchy series

A series \(\sum x_n\) is said to be *Cauchy* or a *Cauchy series*, if the sequence of partial sums \(\{ s_n \}\) is a Cauchy sequence.

A sequence of real numbers converges if and only if it is Cauchy. Therefore a series is convergent if and only if it is Cauchy.

The series \(\sum x_n\) is Cauchy if for every \(\epsilon > 0\), there exists an \(M \in {\mathbb{N}}\), such that for every \(n \geq M\) and \(k \geq M\) we have \[\left\lvert { \left( \sum_{j=1}^k x_j \right) - \left( \sum_{j=1}^n x_j \right) } \right\rvert < \epsilon .\] Without loss of generality we assume \(n < k\). Then we write \[\left\lvert { \left( \sum_{j=1}^k x_j \right) - \left( \sum_{j=1}^n x_j \right) } \right\rvert = \left\lvert { \sum_{j={n+1}}^k x_j } \right\rvert < \epsilon .\] We have proved the following simple proposition.

[prop:cachyser] The series \(\sum x_n\) is Cauchy if for every \(\epsilon > 0\), there exists an \(M \in {\mathbb{N}}\) such that for every \(n \geq M\) and every \(k > n\) we have \[\left\lvert { \sum_{j={n+1}}^k x_j } \right\rvert < \epsilon .\]

#### Basic properties

Let \(\sum x_n\) be a convergent series. Then the sequence \(\{ x_n \}\) is convergent and \[\lim_{n\to\infty} x_n = 0.\]

Let \(\epsilon > 0\) be given. As \(\sum x_n\) is convergent, it is Cauchy. Thus we find an \(M\) such that for every \(n \geq M\) we have \[\epsilon > \left\lvert { \sum_{j={n+1}}^{n+1} x_j } \right\rvert = \left\lvert { x_{n+1} } \right\rvert .\] Hence for every \(n \geq M+1\) we have \(\left\lvert {x_{n}} \right\rvert < \epsilon\).

So if a series converges, the terms of the series go to zero. The implication, however, goes only one way. Let us give an example.

[example:harmonicseries] The series \(\sum \frac{1}{n}\) diverges (despite the fact that \(\lim \frac{1}{n} = 0\)). This is the famous *harmonic series* ^{12}.

Proof: We will show that the sequence of partial sums is unbounded, and hence cannot converge. Write the partial sums \(s_n\) for \(n = 2^k\) as: \[\begin{aligned} s_1 & = 1 , \\ s_2 & = \left( 1 \right) + \left( \frac{1}{2} \right) , \\ s_4 & = \left( 1 \right) + \left( \frac{1}{2} \right) + \left( \frac{1}{3} + \frac{1}{4} \right) , \\ s_8 & = \left( 1 \right) + \left( \frac{1}{2} \right) + \left( \frac{1}{3} + \frac{1}{4} \right) + \left( \frac{1}{5} + \frac{1}{6} + \frac{1}{7} + \frac{1}{8} \right) , \\ & ~~ \vdots \\ s_{2^k} & = 1 + \sum_{j=1}^k \left( \sum_{m=2^{j-1}+1}^{2^j} \frac{1}{m} \right) .\end{aligned}\] We note that \(\nicefrac{1}{3} + \nicefrac{1}{4} \geq \nicefrac{1}{4} + \nicefrac{1}{4} = \nicefrac{1}{2}\) and \(\nicefrac{1}{5} + \nicefrac{1}{6} + \nicefrac{1}{7} + \nicefrac{1}{8} \geq \nicefrac{1}{8} + \nicefrac{1}{8} + \nicefrac{1}{8} + \nicefrac{1}{8} = \nicefrac{1}{2}\). More generally \[\sum_{m=2^{k-1}+1}^{2^k} \frac{1}{m} \geq \sum_{m=2^{k-1}+1}^{2^k} \frac{1}{2^k} = (2^{k-1}) \frac{1}{2^k} = \frac{1}{2} .\] Therefore \[s_{2^k} = 1 + \sum_{j=1}^k \left( \sum_{m=2^{k-1}+1}^{2^k} \frac{1}{m} \right) \geq 1 + \sum_{j=1}^k \frac{1}{2} = 1 + \frac{k}{2} .\] As \(\{ \frac{k}{2} \}\) is unbounded by the , that means that \(\{ s_{2^k} \}\) is unbounded, and therefore \(\{ s_n \}\) is unbounded. Hence \(\{ s_n \}\) diverges, and consequently \(\sum \frac{1}{n}\) diverges.

Convergent series are linear. That is, we can multiply them by constants and add them and these operations are done term by term.

Let \(\alpha \in {\mathbb{R}}\) and \(\sum x_n\) and \(\sum y_n\) be convergent series. Then

- \(\sum \alpha x_n\) is a convergent series and \[\sum_{n=1}^\infty \alpha x_n = \alpha \sum_{n=1}^\infty x_n .\]
- \(\sum ( x_n + y_n )\) is a convergent series and \[\sum_{n=1}^\infty ( x_n + y_n ) = \left( \sum_{n=1}^\infty x_n \right) + \left( \sum_{n=1}^\infty y_n \right) .\]

For the first item, we simply write the \(k\)th partial sum \[\sum_{n=1}^k \alpha x_n = \alpha \left( \sum_{n=1}^k x_n \right) .\] We look at the right-hand side and note that the constant multiple of a convergent sequence is convergent. Hence, we simply take the limit of both sides to obtain the result.

For the second item we also look at the \(k\)th partial sum \[\sum_{n=1}^k ( x_n + y_n ) = \left( \sum_{n=1}^k x_n \right) + \left( \sum_{n=1}^k y_n \right) .\] We look at the right-hand side and note that the sum of convergent sequences is convergent. Hence, we simply take the limit of both sides to obtain the proposition.

Note that multiplying series is not as simple as adding, see the next section. It is not true, of course, that we can multiply term by term, since that strategy does not work even for finite sums. For example, \((a+b)(c+d) \not= ac+bd\).

#### Absolute convergence

Since monotone sequences are easier to work with than arbitrary sequences, it is generally easier to work with series \(\sum x_n\) where \(x_n \geq 0\) for all \(n\). Then the sequence of partial sums is monotone increasing and converges if it is bounded from above. Let us formalize this statement as a proposition.

If \(x_n \geq 0\) for all \(n\), then \(\sum x_n\) converges if and only if the sequence of partial sums is bounded from above.

As the limit of a monotone increasing sequence is the supremum, have the inequality \[\sum_{n=1}^k x_n \leq \sum_{n=1}^\infty x_n .\]

The following criterion often gives a convenient way to test for convergence of a series.

A series \(\sum x_n\) *converges absolutely* if the series \(\sum \left\lvert {x_n} \right\rvert\) converges. If a series converges, but does not converge absolutely, we say it is *conditionally convergent*.

If the series \(\sum x_n\) converges absolutely, then it converges.

A series is convergent if and only if it is Cauchy. Hence suppose \(\sum \left\lvert {x_n} \right\rvert\) is Cauchy. That is, for every \(\epsilon > 0\), there exists an \(M\) such that for all \(k \geq M\) and \(n > k\) we have \[\sum_{j=k+1}^n \left\lvert {x_j} \right\rvert = \left\lvert { \sum_{j=k+1}^n \left\lvert {x_j} \right\rvert } \right\rvert < \epsilon .\] We apply the triangle inequality for a finite sum to obtain \[\left\lvert { \sum_{j=k+1}^n x_j } \right\rvert \leq \sum_{j=k+1}^n \left\lvert {x_j} \right\rvert < \epsilon .\] Hence \(\sum x_n\) is Cauchy and therefore it converges.

Of course, if \(\sum x_n\) converges absolutely, the limits of \(\sum x_n\) and \(\sum \left\lvert {x_n} \right\rvert\) are different. Computing one does not help us compute the other.

Absolutely convergent series have many wonderful properties. For example, absolutely convergent series can be rearranged arbitrarily, or we can multiply such series together easily. Conditionally convergent series on the other hand do not often behave as one would expect. See the next section.

We leave as an exercise to show that \[\sum_{n=1}^\infty \frac{{(-1)}^n}{n}\] converges, although the reader should finish this section before trying. On the other hand we proved \[\sum_{n=1}^\infty \frac{1}{n}\] diverges. Therefore \(\sum \frac{{(-1)}^n}{n}\) is a conditionally convergent subsequence.

#### Comparison test and the p-series

We have noted above that for a series to converge the terms not only have to go to zero, but they have to go to zero “fast enough.” If we know about convergence of a certain series we can use the following comparison test to see if the terms of another series go to zero “fast enough.”

Let \(\sum x_n\) and \(\sum y_n\) be series such that \(0 \leq x_n \leq y_n\) for all \(n \in {\mathbb{N}}\).

- If \(\sum y_n\) converges, then so does \(\sum x_n\).
- If \(\sum x_n\) diverges, then so does \(\sum y_n\).

Since the terms of the series are all nonnegative, the sequences of partial sums are both monotone increasing. Since \(x_n \leq y_n\) for all \(n\), the partial sums satisfy for all \(k\) \[\label{comptest:eq} \sum_{n=1}^k x_n \leq \sum_{n=1}^k y_n .\] If the series \(\sum y_n\) converges the partial sums for the series are bounded. Therefore the right-hand side of [comptest:eq] is bounded for all \(k\). Hence the partial sums for \(\sum x_n\) are also bounded. Since the partial sums are a monotone increasing sequence they are convergent. The first item is thus proved.

On the other hand if \(\sum x_n\) diverges, the sequence of partial sums must be unbounded since it is monotone increasing. That is, the partial sums for \(\sum x_n\) are eventually bigger than any real number. Putting this together with [comptest:eq] we see that for any \(B \in {\mathbb{R}}\), there is a \(k\) such that \[B \leq \sum_{n=1}^k x_n \leq \sum_{n=1}^k y_n .\] Hence the partial sums for \(\sum y_n\) are also unbounded, and \(\sum y_n\) also diverges.

A useful series to use with the comparison test is the \(p\)-series.

For \(p \in {\mathbb{R}}\), the series \[\sum_{n=1}^\infty \frac{1}{n^p}\] converges if and only if \(p > 1\).

First suppose \(p \leq 1\). As \(n \geq 1\), we have \(\frac{1}{n^p} \geq \frac{1}{n}\). Since \(\sum \frac{1}{n}\) diverges, we see that the \(\sum \frac{1}{n^p}\) must diverge for all \(p \leq 1\) by the comparison test.

Now suppose \(p > 1\). We proceed in a similar fashion as we did in the case of the harmonic series, but instead of showing that the sequence of partial sums is unbounded we show that it is bounded. Since the terms of the series are positive, the sequence of partial sums is monotone increasing and will converge if we show that it is bounded above. Let \(s_n\) denote the \(n\)th partial sum. \[\begin{aligned} s_1 & = 1 , \\ s_3 & = \left( 1 \right) + \left( \frac{1}{2^p} + \frac{1}{3^p} \right) , \\ s_7 & = \left( 1 \right) + \left( \frac{1}{2^p} + \frac{1}{3^p} \right) + \left( \frac{1}{4^p} + \frac{1}{5^p} + \frac{1}{6^p} + \frac{1}{7^p} \right) , \\ & ~~ \vdots \\ s_{2^k - 1} &= 1 + \sum_{j=1}^{k-1} \left( \sum_{m=2^j}^{2^{j+1}-1} \frac{1}{m^p} \right) .\end{aligned}\] Instead of estimating from below, we estimate from above. In particular, as \(p\) is positive, then \(2^p < 3^p\), and hence \(\frac{1}{2^p} + \frac{1}{3^p} < \frac{1}{2^p} + \frac{1}{2^p}\). Similarly \(\frac{1}{4^p} + \frac{1}{5^p} + \frac{1}{6^p} + \frac{1}{7^p} < \frac{1}{4^p} + \frac{1}{4^p} + \frac{1}{4^p} + \frac{1}{4^p}\). Therefore \[\begin{split} s_{2^k-1} & = 1+ \sum_{j=1}^k \left( \sum_{m=2^{j}}^{2^{j+1}-1} \frac{1}{m^p} \right) \\ & < 1+ \sum_{j=1}^k \left( \sum_{m=2^{j}}^{2^{j+1}-1} \frac{1}

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[13]/div/div[5]/p[8]/span, line 1, column 3 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Note that neither the \(p\)-series test nor the comparison test tell us what the sum converges to. They only tell us that a limit of the partial sums exists. For example, while we know that \(\sum \nicefrac{1}{n^2}\) converges it is far harder to find^{13} that the limit is \(\nicefrac{\pi^2}{6}\). If we treat \(\sum \nicefrac{1}{n^p}\) as a function of \(p\), we get the so-called Riemann \(\zeta\) function. Understanding the behavior of this function contains one of the most famous unsolved problems in mathematics today and has applications in seemingly unrelated areas such as modern cryptography.

The series \(\sum \frac{1}{n^2+1}\) converges.

Proof: First note that \(\frac{1}{n^2+1} < \frac{1}{n^2}\) for all \(n \in {\mathbb{N}}\). Note that \(\sum \frac{1}{n^2}\) converges by the \(p\)-series test. Therefore, by the comparison test, \(\sum \frac{1}{n^2+1}\) converges.

#### Ratio test

Let \(\sum x_n\) be a series such that \[L := \lim_{n\to\infty} \frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert}\] exists. Then

- If \(L < 1\), then \(\sum x_n\) converges absolutely.
- If \(L > 1\), then \(\sum x_n\) diverges.

From we note that if \(L > 1\), then \(x_n\) diverges. Since it is a necessary condition for the convergence of series that the terms go to zero, we know that \(\sum x_n\) must diverge.

Thus suppose \(L < 1\). We will argue that \(\sum \left\lvert {x_n} \right\rvert\) must converge. The proof is similar to that of . Of course \(L \geq 0\). Pick \(r\) such that \(L < r < 1\). As \(r-L > 0\), there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) \[\left\lvert {\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} - L} \right\rvert < r-L .\] Therefore, \[\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} < r .\] For \(n > M\) (that is for \(n \geq M+1\)) write \[\left\lvert {x_n} \right\rvert = \left\lvert {x_M} \right\rvert \frac{\left\lvert {x_{M+1}} \right\rvert}{\left\lvert {x_{M}} \right\rvert} \frac{\left\lvert {x_{M+2}} \right\rvert}{\left\lvert {x_{M+1}} \right\rvert} \cdots \frac{\left\lvert {x_{n}} \right\rvert}{\left\lvert {x_{n-1}} \right\rvert} < \left\lvert {x_M} \right\rvert r r \cdots r = \left\lvert {x_M} \right\rvert r^{n-M} = (\left\lvert {x_M} \right\rvert r^{-M}) r^n .\] For \(k > M\) we write the partial sum as \[\begin{split} \sum_{n=1}^k \left\lvert {x_n} \right\rvert & = \left(\sum_{n=1}^{M} \left\lvert {x_n} \right\rvert \right) + \left(\sum_{n=M+1}^{k} \left\lvert {x_n} \right\rvert \right) \\ & \leq \left(\sum_{n=1}^{M} \left\lvert {x_n} \right\rvert \right) + \left(\sum_{n=M+1}^{k} (\left\lvert {x_M} \right\rvert r^{-M}) r^n \right) \\ & \leq \left(\sum_{n=1}^{M} \left\lvert {x_n} \right\rvert \right) + (\left\lvert {x_M} \right\rvert r^{-M}) \left( \sum_{n=M+1}^{k} r^n \right) . \end{split}\] As \(0 < r < 1\) the geometric series \(\sum_{n=0}^{\infty} r^n\) converges, so \(\sum_{n=M+1}^{\infty} r^n\) converges as well (why?). We take the limit as \(k\) goes to infinity on the right-hand side above to obtain \[\begin{split} \sum_{n=1}^k \left\lvert {x_n} \right\rvert & \leq \left(\sum_{n=1}^{M} \left\lvert {x_n} \right\rvert \right) + (\left\lvert {x_M} \right\rvert r^{-M}) \left( \sum_{n=M+1}^{k} r^n \right) \\ & \leq \left(\sum_{n=1}^{M} \left\lvert {x_n} \right\rvert \right) + (\left\lvert {x_M} \right\rvert r^{-M}) \left( \sum_{n=M+1}^{\infty} r^n \right) . \end{split}\] The right-hand side is a number that does not depend on \(n\). Hence the sequence of partial sums of \(\sum \left\lvert {x_n} \right\rvert\) is bounded and \(\sum \left\lvert {x_n} \right\rvert\) is convergent. Thus \(\sum x_n\) is absolutely convergent.

The series \[\sum_{n=1}^\infty \frac{2^n}{n!}\] converges absolutely.

Proof: We write \[\lim_{n\to\infty} \frac{2^{(n+1)}/(n+1)!}{2^n / n!} = \lim_{n\to\infty} \frac{2}{n+1} = 0 .\] Therefore, the series converges absolutely by the ratio test.

## Exercises

For \(r \not= 1\), prove \[\sum_{k=0}^{n-1} r^k = \frac{1-r^n}{1-r} .\] Hint: Let \(s := \sum_{k=0}^{n-1} r^k\), then compute \(s(1-r) = s-rs\), and solve for \(s\).

[geometric:exr] Prove that for \(-1 < r < 1\) we have \[\sum_{n=0}^\infty r^n = \frac{1}{1-r} .\] Hint: Use the previous exercise.

Decide the convergence or divergence of the following series.

a) \(\displaystyle \sum_{n=1}^\infty \frac{3}{9n+1}\) | b) \(\displaystyle \sum_{n=1}^\infty \frac{1}{2n-1}\) | c) \(\displaystyle \sum_{n=1}^\infty \frac{{(-1)}^n}{n^2}\) | |||

d) \(\displaystyle \sum_{n=1}^\infty \frac{1}{n(n+1)}\) | e) \(\displaystyle \sum_{n=1}^\infty n e^{-n^2}\) |

- Prove that if \(\displaystyle \sum_{n=1}^\infty x_n\) converges, then \(\displaystyle \sum_{n=1}^\infty ( x_{2n} + x_{2n+1} )\) also converges.
- Find an explicit example where the converse does not hold.

For \(j=1,2,\ldots,n\), let \(\{ x_{j,k} \}_{k=1}^\infty\) denote \(n\) sequences. Suppose that for each \(j\) \[\sum_{k=1}^\infty x_{j,k}\] is convergent. Then show \[\sum_{j=1}^n \left( \sum_{k=1}^\infty x_{j,k} \right) = \sum_{k=1}^\infty \left( \sum_{j=1}^n x_{j,k} \right) .\]

Prove the following stronger version of the ratio test: Let \(\sum x_n\) be a series.

- If there is an \(N\) and a \(\rho < 1\) such that for all \(n \geq N\) we have \(\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} < \rho\), then the series converges absolutely.
- If there is an \(N\) such that for all \(n \geq N\) we have \(\frac{\left\lvert {x_{n+1}} \right\rvert}{\left\lvert {x_n} \right\rvert} \geq 1\), then the series diverges.

Let \(\{ x_n \}\) be a decreasing sequence such that \(\sum x_n\) converges. Show that \(\displaystyle \lim_{n\to\infty} n x_n = 0\).

Show that \(\displaystyle \sum_{n=1}^\infty \frac{{(-1)}^n}{n}\) converges. Hint: consider the sum of two subsequent entries.

- Prove that if \(\sum x_n\) and \(\sum y_n\) converge absolutely, then \(\sum x_ny_n\) converges absolutely.
- Find an explicit example where the converse does not hold.
- Find an explicit example where all three series are absolutely convergent, are not just finite sums, and \((\sum x_n)(\sum y_n) \not= \sum x_ny_n\). That is, show that series are not multiplied term-by-term.

Prove the triangle inequality for series, that is if \(\sum x_n\) converges absolutely then \[\left\lvert {\sum_{n=1}^\infty x_n} \right\rvert \leq \sum_{n=1}^\infty \left\lvert {x_n} \right\rvert .\]

Prove the *limit comparison test*. That is, prove that if \(a_n > 0\) and \(b_n > 0\) for all \(n\), and \[0 < \lim_{n\to\infty} \frac{a_n}{b_n} < \infty ,\] then either \(\sum a_n\) and \(\sum b_n\) both converge of both diverge.

[exercise:badnocauchy] Let \(x_n = \sum_{j=1}^n \nicefrac{1}{j}\). Show that for every \(k\) we have \(\displaystyle \lim_{n\to\infty} \left\lvert {x_{n+k}-x_n} \right\rvert = 0\), yet \(\{ x_n \}\) is not Cauchy.

Let \(s_k\) be the \(k\)th partial sum of \(\sum x_n\).

a) Suppose that there exists a \(m \in {\mathbb{N}}\) such that \(\displaystyle \lim_{k\to\infty} s_{mk}\) exists and \(\lim\, x_n = 0\). Show that \(\sum x_n\) converges.

b) Find an example where \(\displaystyle \lim_{k\to\infty} s_{2k}\) exists and \(\lim\, x_n \not= 0\) (and therefore \(\sum x_n\) diverges).

c) (Challenging) Find an example where \(\lim\, x_n = 0\), and there exists a subsequence \(\{ s_{k_j} \}\) such that \(\displaystyle \lim_{j\to\infty} s_{k_j}\) exists, but \(\sum x_n\) still diverges.

### More on series

*Note: up to 2–3 lectures (optional, can safely be skipped or covered partially)*

#### Root test

We have seen the ratio test before. There is one more similar test called the *root test*. In fact, the proof of this test is similar and somewhat easier.

Let \(\sum x_n\) be a series and let \[L := \limsup_{n\to\infty} \, {\left\lvert {x_n} \right\rvert}^{1/n} .\] Then

- If \(L < 1\) then \(\sum x_n\) converges absolutely.
- If \(L > 1\) then \(\sum x_n\) diverges.

If \(L > 1\), then there exists a subsequence \(\{ x_{n_k} \}\) such that \(L = \lim_{k\to\infty} \, {\left\lvert {x_{n_k}} \right\rvert}^{1/n_k}\). Let \(r\) be such that \(L > r > 1\). There exists an \(M\) such that for all \(k \geq M\), we have \({\left\lvert {x_{n_k}} \right\rvert}^{1/n_k} > r > 1\), or in other words \(\left\lvert {x_{n_k}} \right\rvert > r^{n_k} > 1\). The subsequence \(\{ \left\lvert {x_{n_k}} \right\rvert \}\), and therefore also \(\{ \left\lvert {x_{n}} \right\rvert \}\), cannot possibly converge to zero, and so the series diverges.

Now suppose \(L < 1\). Pick \(r\) such that \(L < r < 1\). By definition of limit supremum, pick \(M\) such that for all \(n \geq M\) we have \[\sup \{ {\left\lvert {x_k} \right\rvert}^{1/k} : k \geq n \} < r .\] Therefore, for all \(n \geq M\) we have \[{\left\lvert {x_n} \right\rvert}^{1/n} < r , \qquad \text{or in other words} \qquad \left\lvert {x_n} \right\rvert < r^n .\] Let \(k > M\) and let us estimate the \(k\)th partial sum \[\sum_{n=1}^k \left\lvert {x_n} \right\rvert = \left( \sum_{n=1}^M \left\lvert {x_n} \right\rvert \right) + \left( \sum_{n=M+1}^k \left\lvert {x_n} \right\rvert \right) \leq \left( \sum_{n=1}^M \left\lvert {x_n} \right\rvert \right) + \left( \sum_{n=M+1}^k r^n \right) .\] As \(0 < r < 1\), the geometric series \(\sum_{n=M+1}^\infty r^n\) converges to \(\frac{r^{M+1}}{1-r}\). As everything is positive we have \[\sum_{n=1}^k \left\lvert {x_n} \right\rvert \leq \left( \sum_{n=1}^M \left\lvert {x_n} \right\rvert \right) + \frac{r^{M+1}}{1-r} .\] Thus the sequence of partial sums of \(\sum \left\lvert {x_n} \right\rvert\) is bounded, and so the series converges. Therefore \(\sum x_n\) converges absolutely.

#### Alternating series test

The tests we have so far only addressed absolute convergence. The following test gives a large supply of conditionally convergent series.

Let \(\{ x_n \}\) be a monotone decreasing sequence of positive real numbers such that \(\lim\, x_n = 0\). Then \[\sum_{n=1}^\infty {(-1)}^n x_n\] converges.

Write \(s_m := \sum_{k=1}^m {(-1)}^k x_k\) be the \(m\)th partial sum. Then write \[s_{2n} = \sum_{k=1}^{2n} {(-1)}^k x_k = (-x_1 + x_2) + \cdots + (-x_{2n-1} + x_{2n}) = \sum_{k=1}^{n} (-x_{2k-1} + x_{2k}) .\] The sequence \(\{ x_k \}\) is decreasing and so \((-x_{2k-1}+x_{2k}) \leq 0\) for all \(k\). Therefore the subsequence \(\{ s_{2n} \}\) of partial sums is a decreasing sequence. Similarly, \((x_{2k}-x_{2k+1}) \geq 0\), and so \[s_{2n} = - x_1 + ( x_2 - x_3 ) + \cdots + ( x_{2n-2} - x_{2n-1} ) + x_{2n} \geq -x_1 .\] The sequence \(\{ s_{2n} \}\) is decreasing and bounded below, so it converges. Let \(a := \lim\, s_{2n}\).

We wish to show that \(\lim\, s_m = a\) (not just for the subsequence). Notice \[s_{2n+1} = s_{2n} + x_{2n+1} .\] Given \(\epsilon > 0\), pick \(M\) such that \(\left\lvert {s_{2n}-a} \right\rvert < \nicefrac{\epsilon}{2}\) whenever \(2n \geq M\). Since \(\lim\, x_n = 0\), we also make \(M\) possibly larger to obtain \(x_{2n+1} < \nicefrac{\epsilon}{2}\) whenever \(2n \geq M\). If \(2n \geq M\), we have \(\left\lvert {s_{2n}-a} \right\rvert < \nicefrac{\epsilon}{2} < \epsilon\), so we just need to check the situation for \(s_{2n+1}\): \[\left\lvert {s_{2n+1}-a} \right\rvert = \left\lvert {s_{2n}-a + x_{2n+1}} \right\rvert \leq \left\lvert {s_{2n}-a} \right\rvert + x_{2n+1} < \nicefrac{\epsilon}{2}+ \nicefrac{\epsilon}{2} = \epsilon . \qedhere\]

In particular, there exist conditionally convergent series where the absolute values of the terms go to zero arbitrarily slowly. For example, \[\sum_{n=1}^\infty \frac{{(-1)}^n}{n^p}\] converges for arbitrarily small \(p > 0\), but it does not converge absolutely when \(p \leq 1\).

#### Rearrangements

Generally, absolutely convergent series behave as we imagine they should. For example, absolutely convergent series can be summed in any order whatsoever. Nothing of the sort holds for conditionally convergent series (see and ).

Take a series \[\sum_{n=1}^\infty x_n .\] Given a bijective function \(\sigma \colon {\mathbb{N}}\to {\mathbb{N}}\), the corresponding rearrangement is the following series: \[\sum_{k=1}^\infty x_{\sigma(k)} .\] We simply sum the series in a different order.

Let \(\sum x_n\) be an absolutely convergent series converging to a number \(x\). Let \(\sigma \colon {\mathbb{N}}\to {\mathbb{N}}\) be a bijection. Then \(\sum x_{\sigma(n)}\) is absolutely convergent and converges to \(x\).

In other words, a rearrangement of an absolutely convergent series converges (absolutely) to the same number.

Let \(\epsilon > 0\) be given. Then take \(M\) to be such that \[\left\lvert {\left(\sum_{n=1}^M x_n \right) - x} \right\rvert < \frac{\epsilon}{2} \qquad \text{and} \qquad \sum_{n=M+1}^\infty \left\lvert {x_n} \right\rvert < \frac{\epsilon}{2} .\] As \(\sigma\) is a bijection, there exists a number \(K\) such that for each \(n \leq M\), there exists \(k \leq K\) such that \(\sigma(k) = n\). In other words \(\{ 1,2,\ldots,M \} \subset \sigma\bigl(\{ 1,2,\ldots,K \} \bigr)\).

Then for any \(N \geq K\), let \(Q := \max \sigma(\{ 1,2,\ldots,K \})\) and compute \[\begin{split} \left\lvert {\left( \sum_{n=1}^N x_{\sigma(n)} \right) - x} \right\rvert & = \left\lvert { \left( \sum_{n=1}^M x_n + \sum_{\substack{n=1\\\sigma(n) > M}}^N x_{\sigma(n)} \right) - x} \right\rvert \\ & \leq \left\lvert { \left( \sum_{n=1}^M x_n \right) - x} \right\rvert + \sum_{\substack{n=1\\\sigma(n) > M}}^N \left\lvert {x_{\sigma(n)}} \right\rvert \\ & \leq \left\lvert { \left( \sum_{n=1}^M x_n \right) - x} \right\rvert + \sum_{n=M+1}^Q \left\lvert {x_{n}} \right\rvert \\ & < \nicefrac{\epsilon}{2} + \nicefrac{\epsilon}{2} = \epsilon . \end{split}\] So \(\sum x_{\sigma(n)}\) converges to \(x\). To see that the convergence is absolute, we apply the above argument to \(\sum \left\lvert {x_n} \right\rvert\) to show that \(\sum \left\lvert {x_{\sigma(n)}} \right\rvert\) converges.

[example:harmonsumanything] Let us show that the alternating harmonic series \(\sum \frac

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[14]/div/div[3]/p[7]/span[2], line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[14]/div/div[3]/p[7]/span[3], line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

By construction \(\sigma \colon {\mathbb{N}}\to {\mathbb{N}}\) is one-to-one. It is also onto, because if we keep adding either odd (resp. even) terms, eventually we will pass \(L\) and switch to the evens (resp. odds). So we switch infinitely many times.

Finally, let \(N\) be the \(N\) where we just pass \(L\) and switch. For example suppose we have just switched from odd to even (so we start subtracting), and let \(N' > N\) be where we first switch back from even to odd. Then \[L + \frac{1}{\sigma(N)} \geq \sum_{n=1}^{N-1} a_{\sigma(n)} > \sum_{n=1}^{N'-1} a_{\sigma(n)} > L- \frac{1}{\sigma(N')}.\] And similarly for switching in the other direction. Therefore, the sum up to \(N'-1\) is within \(\frac{1}{\min \{ \sigma(N), \sigma(N') \}}\) of \(L\). As we switch infinitely many times we obtain that \(\sigma(N) \to \infty\) and \(\sigma(N') \to \infty\), and hence \[\sum_{n=1}^\infty a_{\sigma(n)} = \sum_{n=1}^\infty \frac

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[14]/div/div[3]/p[9]/span, line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Here is an example to illustrate the proof. Suppose \(L=1.2\), then the order is \[1+\nicefrac{1}{3}-\nicefrac{1}{2}+\nicefrac{1}{5}+\nicefrac{1}{7}+\nicefrac{1}{9}-\nicefrac{1}{4}+\nicefrac{1}{11}+\nicefrac{1}{13}-\nicefrac{1}{6} +\nicefrac{1}{15}+\nicefrac{1}{17}+\nicefrac{1}{19} - \nicefrac{1}{8} + \cdots .\] At this point we are no more than \(\nicefrac{1}{8}\) from the limit.

#### Multiplication of series

As we have already mentioned, multiplication of series is somewhat harder than addition. If we have that at least one of the series converges absolutely, than we can use the following theorem. For this result it is convenient to start the series at 0, rather than at 1.

Suppose \(\sum_{n=0}^\infty a_n\) and \(\sum_{n=0}^\infty b_n\) are two convergent series, converging to \(A\) and \(B\) respectively. If at least one of the series converges absolutely, then the series \(\sum_{n=0}^\infty c_n\) where \[c_n = a_0 b_n + a_1 b_{n-1} + \cdots + a_n b_0 = \sum_{j=0}^n a_j b_{n-j} ,\] converges to \(AB\).

The series \(\sum c_n\) is called the *Cauchy product* of \(\sum a_n\) and \(\sum b_n\).

Suppose \(\sum a_n\) converges absolutely, and let \(\epsilon > 0\) be given. In this proof instead of picking complicated estimates just to make the final estimate come out as less than \(\epsilon\), let us simply obtain an estimate that depends on \(\epsilon\) and can be made arbitrarily small.

Write \[A_m := \sum_{n=0}^m a_n , \qquad B_m := \sum_{n=0}^m b_n .\] We rearrange the \(m\)th partial sum of \(\sum c_n\): \[\begin{split} \left\lvert {\left(\sum_{n=0}^m c_n \right) - AB} \right\rvert & = \left\lvert {\left( \sum_{n=0}^m \sum_{j=0}^n a_j b_{n-j} \right) - AB} \right\rvert \\ & = \left\lvert {\left( \sum_{n=0}^m B_n a_{m-n} \right) - AB} \right\rvert \\ & = \left\lvert {\left( \sum_{n=0}^m ( B_n - B ) a_{m-n} \right) + B A_m - AB} \right\rvert \\ & \leq \left( \sum_{n=0}^m \left\lvert { B_n - B } \right\rvert \left\lvert {a_{m-n}} \right\rvert \right) + \left\lvert {B} \right\rvert\left\lvert {A_m - A} \right\rvert \end{split}\] We can surely make the second term on the right hand side go to zero. The trick is to handle the first term. Pick \(K\) such that for all \(m \geq K\) we have \(\left\lvert {A_m - A} \right\rvert < \epsilon\) and also \(\left\lvert {B_m - B} \right\rvert < \epsilon\). Finally, as \(\sum a_n\) converges absolutely, make sure that \(K\) is large enough such that for all \(m \geq K\), \[\sum_{n=K}^m \left\lvert {a_n} \right\rvert < \epsilon .\] As \(\sum b_n\) converges, then we have that \(B_{\text{max}} := \sup \{ \left\lvert { B_n - B } \right\rvert : n = 0,1,2,\ldots \}\) is finite. Take \(m \geq 2K\), then in particular \(m-K+1 > K\). So \[\begin{split} %\left( \sum_{n=0}^m \left\lvert { B_n - B } \right\rvert \left\lvert {a_{m-n}} \right\rvert %\right) & = \left( \sum_{n=0}^{m-K} \left\lvert { B_n - B } \right\rvert \left\lvert {a_{m-n}} \right\rvert \right) + \left( \sum_{n=m-K+1}^m \left\lvert { B_n - B } \right\rvert \left\lvert {a_{m-n}} \right\rvert \right) \\ & < \left( \sum_{n=K}^m \left\lvert {a_{n}} \right\rvert \right) B_{\text{max}} + \left( \sum_{n=0}^{K-1} \epsilon \left\lvert {a_{n}} \right\rvert \right) \\ & < \epsilon B_{\text{max}} + \epsilon \left( \sum_{n=0}^\infty \left\lvert {a_{n}} \right\rvert \right) . \end{split}\] Therefore, for \(m \geq 2K\) we have \[\begin{split} \left\lvert {\left(\sum_{n=0}^m c_n \right) - AB} \right\rvert & \leq \left( \sum_{n=0}^m \left\lvert { B_n - B } \right\rvert \left\lvert {a_{m-n}} \right\rvert \right) + \left\lvert {B} \right\rvert\left\lvert {A_m - A} \right\rvert \\ & < \epsilon B_{\text{max}} + \epsilon \left( \sum_{n=0}^\infty \left\lvert {a_{n}} \right\rvert \right) + \left\lvert {B} \right\rvert\epsilon = \epsilon \left( B_{\text{max}} + \left( \sum_{n=0}^\infty \left\lvert {a_{n}} \right\rvert \right) + \left\lvert {B} \right\rvert \right) . \end{split}\] The expression in the parenthesis on the right hand side is a fixed number. Hence, we can make the right hand side arbitrarily small by picking a small enough \(\epsilon> 0\). So \(\sum_{n=0}^\infty c_n\) converges to \(AB\).

If both series are only conditionally convergent, the Cauchy product series need not even converge. Suppose we take \(a_n = b_n = {(-1)}^n \frac{1}{\sqrt{n+1}}\). The series \(\sum_{n=0}^\infty a_n = \sum_{n=0}^\infty b_n\) converges by the alternating series test, however, it does not converge absolutely as can be seen from the \(p\)-test. Let us look at the Cauchy product. \[c_n = {(-1)}^n \left( \frac{1}{\sqrt{n+1}} + \frac{1}{\sqrt{2n}} + \frac{1}{\sqrt{3(n-1)}} + \cdots + %\frac{1}{\sqrt{2n}} + \frac{1}{\sqrt{n+1}} \right) = {(-1)}^n \sum_{j=0}^n \frac{1}{\sqrt{(j+1)(n-j+1)}} .\] Therefore \[\left\lvert {c_n} \right\rvert = \sum_{j=0}^n \frac{1}{\sqrt{(j+1)(n-j+1)}} \geq \sum_{j=0}^n \frac{1}{\sqrt{(n+1)(n+1)}} = 1 .\] The terms do not go to zero and hence \(\sum c_n\) cannot converge.

#### Power series

Fix \(x_0 \in {\mathbb{R}}\). A *power series* about \(x_0\) is a series of the form \[\sum_{n=0}^\infty a_n {(x-x_0)}^n .\] A power series is really a function of \(x\), and many important functions in analysis can be written as a power series.

We say that a power series is *convergent* if there is at least one \(x \not= x_0\) that makes the series converge. Note that it is trivial to see that if \(x=x_0\) then the series always converges since all terms except the first are zero. If the series does not converge for any point \(x \not= x_0\), we say that the series is *divergent*.

[ps:expex] The series \[\sum_{n=0}^\infty \frac{1}{n!} x^n\] is absolutely convergent for all \(x \in {\mathbb{R}}\). This can be seen using the ratio test: For any \(x\) notice that \[\lim_{n \to \infty} \frac{\bigl(1/(n+1)!\bigr) \, x^{n+1}}{(1/n!) \, x^{n}} = \lim_{n \to \infty} \frac{x}{n+1} = 0.\] In fact, you may recall from calculus that this series converges to \(e^x\).

[ps:1kex] The series \[\sum_{n=1}^\infty \frac{1}{n} x^n\] converges absolutely for all \(x \in (-1,1)\) via the ratio test: \[\lim_{n \to \infty} \left\lvert { \frac{\bigl(1/(n+1) \bigr) \, x^{n+1}}{(1/n) \, x^{n}} } \right\rvert = \lim_{n \to \infty} \left\lvert {x} \right\rvert \frac{n}{n+1} = \left\lvert {x} \right\rvert < 1 .\] It converges at \(x=-1\), as \(\sum_{n=1}^\infty \frac{{(-1)}^n}{n}\) converges by the alternating series test. But the power series does not converge absolutely at \(x=-1\), because \(\sum_{n=1}^\infty \frac{1}{n}\) does not converge. The series diverges at \(x=1\). When \(\left\lvert {x} \right\rvert > 1\), then the series diverges via the ratio test.

[ps:divergeex] The series \[\sum_{n=1}^\infty n^n x^n\] diverges for all \(x \not= 0\). Let us apply the root test \[\limsup_{n\to\infty} \, \left\lvert {n^n x^n} \right\rvert^{1/n} = \limsup_{n\to\infty} \, n \left\lvert {x} \right\rvert = \infty .\] Therefore the series diverges for all \(x \not= 0\).

In fact, convergence of power series in general always works analogously to one of the three examples above.

Let \(\sum a_n {(x-x_0)}^n\) be a power series. If the series is convergent, then either it converges at all \(x \in {\mathbb{R}}\), or there exists a number \(\rho\), such that the series converges absolutely on the interval \((x_0-\rho,x_0+\rho)\) and diverges when \(x < x_0-\rho\) or \(x > x_0+\rho\).

The number \(\rho\) is called the *radius of convergence* of the power series. We write \(\rho = \infty\) if the series converges for all \(x\), and we write \(\rho = 0\) if the series is divergent. See . In the radius of convergence is \(\rho=1\). In the radius of convergence is \(\rho=\infty\), and in the radius of convergence is \(\rho=0\).

Write \[R := \limsup_{n\to\infty} \, {\left\lvert {a_n} \right\rvert}^{1/n} .\] We use the root test to prove the proposition: \[L = \limsup_{n\to\infty} \, {\left\lvert {a_n{(x-x_0)}^n} \right\rvert}^{1/n} = \left\lvert {x-x_0} \right\rvert \limsup_{n\to\infty} \, {\left\lvert {a_n} \right\rvert}^{1/n} = \left\lvert {x-x_0} \right\rvert R .\] In particular if \(R = \infty\), then \(L=\infty\) for any \(x \not= x_0\), and the series diverges by the root test. On the other hand if \(R = 0\), then \(L=0\) for any \(x\), and the series converges absolutely for all \(x\).

Suppose \(0 < R < \infty\). The series converges absolutely if \(1 > L = R \left\lvert {x-x_0} \right\rvert\), or in other words when \[\left\lvert {x-x_0} \right\rvert < \nicefrac{1}{R} .\] The series diverges when \(1 < L = R \left\lvert {x-x_0} \right\rvert\), or \[\left\lvert {x-x_0} \right\rvert > \nicefrac{1}{R} .\] Letting \(\rho = \nicefrac{1}{R}\) completes the proof.

It may be useful to restate what we have learned in the proof as a separate proposition.

Let \(\sum a_n {(x-x_0)}^n\) be a power series, and let \[R := \limsup_{n\to\infty} \, {\left\lvert {a_n} \right\rvert}^{1/n} .\] If \(R = \infty\), the power series is divergent. If \(R=0\), then the power series converges everywhere. Otherwise the radius of convergence \(\rho = \nicefrac{1}{R}\).

Often, radius of convergence is written as \(\rho = \nicefrac{1}{R}\) in all three cases, with the obvious understanding of what \(\rho\) should be if \(R = 0\) or \(R = \infty\).

Convergent power series can be added and multiplied together, and multiplied by constants. The proposition has an easy proof using what we know about series in general, and power series in particular. We leave the proof to the reader.

Let \(\sum_{n=0}^\infty a_n {(x-x_0)}^n\) and \(\sum_{n=0}^\infty b_n {(x-x_0)}^n\) be two convergent power series with radius of convergence at least \(\rho > 0\) and \(\alpha \in {\mathbb{R}}\). Then for all \(x\) such that \(\left\lvert {x-x_0} \right\rvert < \rho\), we have \[\left(\sum_{n=0}^\infty a_n {(x-x_0)}^n\right) + \left(\sum_{n=0}^\infty b_n {(x-x_0)}^n\right) = \sum_{n=0}^\infty (a_n+b_n) {(x-x_0)}^n ,\] \[\alpha \left(\sum_{n=0}^\infty a_n {(x-x_0)}^n\right) = \sum_{n=0}^\infty \alpha a_n {(x-x_0)}^n ,\] and \[\left(\sum_{n=0}^\infty a_n {(x-x_0)}^n\right) \, \left(\sum_{n=0}^\infty b_n {(x-x_0)}^n\right) = \sum_{n=0}^\infty c_n {(x-x_0)}^n ,\] where \(c_n = a_0b_n + a_1 b_{n-1} + \cdots + a_n b_0\).

That is, after performing the algebraic operations, the radius of convergence of the resulting series is at least \(\rho\). For all \(x\) with \(\left\lvert {x-x_0} \right\rvert < \rho\), we have two convergent series so their term by term addition and multiplication by constants follows by what we learned in the last section. For multiplication of two power series, the series are absolutely convergent inside the radius of convergence and that is why for those \(x\) we can apply Mertens’ theorem. Note that after applying an algebraic operation the radius of convergence could increase. See the exercises.

Let us look at some examples of power series. Polynomials are simply finite power series. That is, a polynomial is a power series where the \(a_n\) are zero for all \(n\) large enough. We expand a polynomial as a power series about any point \(x_0\) by writing the polynomial as a polynomial in \((x-x_0)\). For example, \(2x^2-3x+4\) as a power series around \(x_0 = 1\) is \[2x^2-3x+4 = 3 + (x-1) + 2{(x-1)}^2 .\]

We can also expand *rational functions*, that is, ratios of polynomials as power series, although we will not completely prove this fact here. Notice that a series for a rational function only defines the function on an interval even if the function is defined elsewhere. For example, for the *geometric series* we have that for \(x \in (-1,1)\) \[\frac{1}{1-x} = \sum_{n=0}^\infty x^n .\] The series diverges when \(\left\lvert {x} \right\rvert > 1\), even though \(\frac{1}{1-x}\) is defined for all \(x \not= 1\).

We can use the geometric series together with rules for addition and multiplication of power series to expand rational functions as power series around \(x_0\), as long as the denominator is not zero at \(x_0\). We state without proof that this is always possible, and we give an example of such a computation using the geometric series.

Let us expand \(\frac{x}{1+2x+x^2}\) as a power series around the origin (\(x_0 = 0\)) and find the radius of convergence.

Write \(1+2x+x^2 = {(1+x)}^2 = {\bigl(1-(-x)\bigr)}^2\), and suppose \(\left\lvert {x} \right\rvert < 1\). Compute \[\begin{split} \frac{x}{1+2x+x^2} &= x \, {\left( \frac{1}{1-(-x)} \right)}^2 \\ &= x \, {\left( \sum_{n=0}^\infty {(-1)}^n x^n \right)}^2 \\ &= x \, \left( \sum_{n=0}^\infty c_n x^n \right) \\ &= \sum_{n=0}^\infty c_n x^{n+1} , \end{split}\] where using the formula for the product of series we obtain, \(c_0 = 1\), \(c_1 = -1 -1 = -2\), \(c_2 = 1+1+1 = 3\), etc…. Therefore we get that for \(\left\lvert {x} \right\rvert < 1\), \[\frac{x}{1+2x+x^2} = \sum_{n=1}^\infty {(-1)}^{n+1} n x^n .\] The radius of convergence is at least 1. We leave it to the reader to verify that the radius of convergence is exactly equal to 1.

You can use the method of partial fractions you know from calculus. For example, to find the power series for \(\frac{x^3+x}{x^2-1}\) at 0, write \[\frac{x^3+x}{x^2-1} = x + \frac{1}{1+x} - \frac{1}{1-x} = x + \sum_{n=0}^\infty {(-1)}^n x^n - \sum_{n=0}^\infty x^n .\]

## Exercises

Decide the convergence or divergence of the following series.

a) \(\displaystyle \sum_{n=1}^\infty \frac{1}{2^{2n+1}}\) | b) \(\displaystyle \sum_{n=1}^\infty \frac{{(-1)}^{n}(n-1)}{n}\) | c) \(\displaystyle \sum_{n=1}^\infty \frac ParseError: EOF expected (click for details) \)Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[15]/table[1]/tbody/tr/td[5]/span, line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14 |
d) \(\displaystyle \sum_{n=1}^\infty \frac{n^n} ParseError: EOF expected (click for details) \)Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[15]/table[1]/tbody/tr/td[7]/span, line 1, column 6 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14 |

Suppose both \(\sum_{n=0}^\infty a_n\) and \(\sum_{n=0}^\infty b_n\) converge absolutely. Show that the product series, \(\sum_{n=0}^\infty c_n\) where \(c_n = a_0 b_n + a_1 b_{n-1} + \cdots + a_n b_0\), also converges absolutely.

[exercise:seriesconvergestoanything] Let \(\sum a_n\) be conditionally convergent. Show that given any number \(x\) there exists a rearrangement of \(\sum a_n\) such that the rearranged series converges to \(x\). Hint: See .

a) Show that the alternating harmonic series \(\sum \frac

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[15]/p[4]/span, line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

For the following power series, find if they are convergent or not, and if so find their radius of convergence.

a) \(\displaystyle \sum_{n=0}^\infty 2^n x^n\) | b) \(\displaystyle \sum_{n=0}^\infty n x^n\) | c) \(\displaystyle \sum_{n=0}^\infty n! \, x^n\) | d) \(\displaystyle \sum_{n=0}^\infty \frac{1}{(2k)!} {(x-10)}^n\) | e) \(\displaystyle \sum_{n=0}^\infty x^{2n}\) | f) \(\displaystyle \sum_{n=0}^\infty n! \, x^{n!}\) |

Suppose \(\sum a_n x^n\) converges for \(x=1\). a) What can you say about the radius of convergence? b) If you further know that at \(x=1\) the convergence is not absolute, what can you say?

Expand \(\dfrac{x}{4-x^2}\) as a power series around \(x_0 = 0\) and compute its radius of convergence.

a) Find an example where the radius of convergence of \(\sum a_n x^n\) and \(\sum b_n x^n\) are 1, but the radius of convergence of the sum of the two series is infinite. b) (Trickier) Find an example where the radius of convergence of \(\sum a_n x^n\) and \(\sum b_n x^n\) are 1, but the radius of convergence of the product of the two series is infinite.

Figure out how to compute the radius of convergence using the ratio test. That is, suppose \(\sum a_n x^n\) is a power series and \(R := \lim \, \frac{\left\lvert {a_{n+1}} \right\rvert}{\left\lvert {a_n} \right\rvert}\) exists or is \(\infty\). Find the radius of convergence and prove your claim.

a) Prove that \(\lim \, n^{1/n} = 1\). Hint: Write \(n^{1/n} = 1+b_n\) and note \(b_n > 0\). Then show that \({(1+b_n)}^n \geq \frac{n(n-1)}{2}b_n^2\) and use this to show that \(\lim \, b_n = 0\). b) Use the result of part a) to show that if \(\sum a_n x^n\) is a convergent power series with radius of convergence \(R\), then \(\sum n a_n x^n\) is also convergent with the same radius of convergence.

There are different notions of summability (convergence) of a series than just the one we have seen. A common one is *Cesàro summability* ^{14}. Let \(\sum a_n\) be a series and let \(s_n\) be the \(n\)th partial sum. The series is said to be Cesàro summable to \(a\) if \[a = \lim_{n\to \infty} \frac{s_1 + s_2 + \cdots + s_n}{n} .\]

a) If \(\sum a_n\) is convergent to \(a\) (in the usual sense), show that \(\sum a_n\) is Cesàro summable to \(a\). b) Show that in the sense of Cesàro \(\sum {(-1)}^n\) is summable to \(\nicefrac{1}{2}\). c) Let \(a_n := k\) when \(n = k^3\) for some \(k \in {\mathbb{N}}\), \(a_n := -k\) when \(n = k^3+1\) for some \(k \in {\mathbb{N}}\), otherwise let \(a_n := 0\). Show that \(\sum a_n\) diverges in the usual sense, (partial sums are unbounded), but it is Cesàro summable to 0 (seems a little paradoxical at first sight).

Show that the monotonicity in the alternating series test is necessary. That is, find a sequence of positive real numbers \(\{ x_n \}\) with \(\lim\, x_n = 0\) but such that \(\sum {(-1)}^n x_n\) diverges.

## Continuous Functions

### Limits of functions

*Note: 2–3 lectures*

Before we define continuity of functions, we need to visit a somewhat more general notion of a limit. That is, given a function \(f \colon S \to {\mathbb{R}}\), we want to see how \(f(x)\) behaves as \(x\) tends to a certain point.

#### Cluster points

First, let us return to a concept we have previously seen in an exercise.

Let \(S \subset {\mathbb{R}}\) be a set. A number \(x \in {\mathbb{R}}\) is called a *cluster point* of \(S\) if for every \(\epsilon > 0\), the set \((x-\epsilon,x+\epsilon) \cap S \setminus \{ x \}\) is not empty.

That is, \(x\) is a cluster point of \(S\) if there are points of \(S\) arbitrarily close to \(x\). Another way of phrasing the definition is to say that \(x\) is a cluster point of \(S\) if for every \(\epsilon > 0\), there exists a \(y \in S\) such that \(y \not= x\) and \(\left\lvert {x - y} \right\rvert < \epsilon\). Note that a cluster point of \(S\) need not lie in \(S\).

Let us see some examples.

- The set \(\{ \nicefrac{1}{n} : n \in {\mathbb{N}}\}\) has a unique cluster point zero.
- The cluster points of the open interval \((0,1)\) are all points in the closed interval \([0,1]\).
- For the set \({\mathbb{Q}}\), the set of cluster points is the whole real line \({\mathbb{R}}\).
- For the set \([0,1) \cup \{ 2 \}\), the set of cluster points is the interval \([0,1]\).
- The set \({\mathbb{N}}\) has no cluster points in \({\mathbb{R}}\).

Let \(S \subset {\mathbb{R}}\). Then \(x \in {\mathbb{R}}\) is a cluster point of \(S\) if and only if there exists a convergent sequence of numbers \(\{ x_n \}\) such that \(x_n \not= x\), \(x_n \in S\), and \(\lim\, x_n = x\).

First suppose \(x\) is a cluster point of \(S\). For any \(n \in {\mathbb{N}}\), we pick \(x_n\) to be an arbitrary point of \((x-\nicefrac{1}{n},x+\nicefrac{1}{n}) \cap S \setminus \{x\}\), which we know is nonempty because \(x\) is a cluster point of \(S\). Then \(x_n\) is within \(\nicefrac{1}{n}\) of \(x\), that is, \[\left\lvert {x-x_n} \right\rvert < \nicefrac{1}{n} .\] As \(\{ \nicefrac{1}{n} \}\) converges to zero, \(\{ x_n \}\) converges to \(x\).

On the other hand, if we start with a sequence of numbers \(\{ x_n \}\) in \(S\) converging to \(x\) such that \(x_n \not= x\) for all \(n\), then for every \(\epsilon > 0\) there is an \(M\) such that in particular \(\left\lvert {x_M - x} \right\rvert < \epsilon\). That is, \(x_M \in (x-\epsilon,x+\epsilon) \cap S \setminus \{x\}\).

#### Limits of functions

If a function \(f\) is defined on a set \(S\) and \(c\) is a cluster point of \(S\), then we can define the limit of \(f(x)\) as \(x\) gets close to \(c\). Do note that it is irrelevant for the definition if \(f\) is defined at \(c\) or not. Furthermore, even if the function is defined at \(c\), the limit of the function as \(x\) goes to \(c\) could very well be different from \(f(c)\).

Let \(f \colon S \to {\mathbb{R}}\) be a function and \(c\) a cluster point of \(S\). Suppose there exists an \(L \in {\mathbb{R}}\) and for every \(\epsilon > 0\), there exists a \(\delta > 0\) such that whenever \(x \in S \setminus \{ c \}\) and \(\left\lvert {x - c} \right\rvert < \delta\), then \[\left\lvert {f(x) - L} \right\rvert < \epsilon .\] In this case we say \(f(x)\) *converges* to \(L\) as \(x\) goes to \(c\). We say \(L\) is the *limit* of \(f(x)\) as \(x\) goes to \(c\). We write \[\lim_{x \to c} f(x) := L ,\] or \[f(x) \to L \quad\text{as}\quad x \to c .\] If no such \(L\) exists, then we say that the limit does not exist or that \(f\) *diverges* at \(c\).

Again the notation and language we are using above assumes the limit is unique even though we have not yet proved that. Let us do that now.

Let \(c\) be a cluster point of \(S \subset {\mathbb{R}}\) and let \(f \colon S \to {\mathbb{R}}\) be a function such that \(f(x)\) converges as \(x\) goes to \(c\). Then the limit of \(f(x)\) as \(x\) goes to \(c\) is unique.

Let \(L_1\) and \(L_2\) be two numbers that both satisfy the definition. Take an \(\epsilon > 0\) and find a \(\delta_1 > 0\) such that \(\left\lvert {f(x)-L_1} \right\rvert < \nicefrac{\epsilon}{2}\) for all \(x \in S \setminus \{c\}\) with \(\left\lvert {x-c} \right\rvert < \delta_1\). Also find \(\delta_2 > 0\) such that \(\left\lvert {f(x)-L_2} \right\rvert < \nicefrac{\epsilon}{2}\) for all \(x \in S \setminus \{c\}\) with \(\left\lvert {x-c} \right\rvert < \delta_2\). Put \(\delta := \min \{ \delta_1, \delta_2 \}\). Suppose \(x \in S\), \(\left\lvert {x-c} \right\rvert < \delta\), and \(x \not= c\). Then \[\left\lvert {L_1 - L_2} \right\rvert = \left\lvert {L_1 - f(x) + f(x) - L_2} \right\rvert \leq \left\lvert {L_1 - f(x)} \right\rvert + \left\lvert {f(x) - L_2} \right\rvert < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon.\] As \(\left\lvert {L_1-L_2} \right\rvert < \epsilon\) for arbitrary \(\epsilon > 0\), then \(L_1 = L_2\).

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be defined as \(f(x) := x^2\). Then \[\lim_{x\to c} f(x) = \lim_{x\to c} x^2 = c^2 .\]

Proof: First let \(c\) be fixed. Let \(\epsilon > 0\) be given. Take \[\delta := \min \left\{ 1 , \, \frac{\epsilon}{2\left\lvert {c} \right\rvert+1} \right\} .\] Take \(x \not= c\) such that \(\left\lvert {x-c} \right\rvert < \delta\). In particular, \(\left\lvert {x-c} \right\rvert < 1\). Then by reverse triangle inequality we get \[\left\lvert {x} \right\rvert-\left\lvert {c} \right\rvert \leq \left\lvert {x-c} \right\rvert < 1 .\] Adding \(2\left\lvert {c} \right\rvert\) to both sides we obtain \(\left\lvert {x} \right\rvert + \left\lvert {c} \right\rvert < 2\left\lvert {c} \right\rvert + 1\). We compute \[\begin{split} \left\lvert {f(x) - c^2} \right\rvert &= \left\lvert {x^2-c^2} \right\rvert \\ &= \left\lvert {(x+c)(x-c)} \right\rvert \\ &= \left\lvert {x+c} \right\rvert\left\lvert {x-c} \right\rvert \\ &\leq (\left\lvert {x} \right\rvert+\left\lvert {c} \right\rvert)\left\lvert {x-c} \right\rvert \\ &< (2\left\lvert {c} \right\rvert+1)\left\lvert {x-c} \right\rvert \\ &< (2\left\lvert {c} \right\rvert+1)\frac{\epsilon}{2\left\lvert {c} \right\rvert+1} = \epsilon . \end{split}\]

Define \(f \colon [0,1) \to {\mathbb{R}}\) by \[f(x) := \begin{cases} x & \text{if $x > 0$} , \\ 1 & \text{if $x = 0$} . \end{cases}\] Then \[\lim_{x\to 0} f(x) = 0 ,\] even though \(f(0) = 1\).

Proof: Let \(\epsilon > 0\) be given. Let \(\delta := \epsilon\). Then for \(x \in [0,1)\), \(x \not= 0\), and \(\left\lvert {x-0} \right\rvert < \delta\) we get \[\left\lvert {f(x) - 0} \right\rvert = \left\lvert {x} \right\rvert < \delta = \epsilon .\]

#### Sequential limits

Let us connect the limit as defined above with limits of sequences.

[seqflimit:lemma] Let \(S \subset {\mathbb{R}}\) and \(c\) be a cluster point of \(S\). Let \(f \colon S \to {\mathbb{R}}\) be a function.

Then \(f(x) \to L\) as \(x \to c\), if and only if for every sequence \(\{ x_n \}\) of numbers such that \(x_n \in S \setminus \{c\}\) for all \(n\), and such that \(\lim\, x_n = c\), we have that the sequence \(\{ f(x_n) \}\) converges to \(L\).

Suppose \(f(x) \to L\) as \(x \to c\), and \(\{ x_n \}\) is a sequence such that \(x_n \in S \setminus \{c\}\) and \(\lim\, x_n = c\). We wish to show that \(\{ f(x_n) \}\) converges to \(L\). Let \(\epsilon > 0\) be given. Find a \(\delta > 0\) such that if \(x \in S \setminus \{c\}\) and \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {f(x) - L} \right\rvert < \epsilon\). As \(\{ x_n \}\) converges to \(c\), find an \(M\) such that for \(n \geq M\) we have that \(\left\lvert {x_n - c} \right\rvert < \delta\). Therefore, for \(n \geq M\), \[\left\lvert {f(x_n) - L} \right\rvert < \epsilon .\] Thus \(\{ f(x_n) \}\) converges to \(L\).

For the other direction, we use proof by contrapositive. Suppose it is not true that \(f(x) \to L\) as \(x \to c\). The negation of the definition is that there exists an \(\epsilon > 0\) such that for every \(\delta > 0\) there exists an \(x \in S \setminus \{c\}\), where \(\left\lvert {x-c} \right\rvert < \delta\) and \(\left\lvert {f(x)-L} \right\rvert \geq \epsilon\).

Let us use \(\nicefrac{1}{n}\) for \(\delta\) in the above statement to construct a sequence \(\{ x_n \}\). We have that there exists an \(\epsilon > 0\) such that for every \(n\), there exists a point \(x_n \in S \setminus \{c\}\), where \(\left\lvert {x_n-c} \right\rvert < \nicefrac{1}{n}\) and \(\left\lvert {f(x_n)-L} \right\rvert \geq \epsilon\). The sequence \(\{ x_n \}\) just constructed converges to \(c\), but the sequence \(\{ f(x_n) \}\) does not converge to \(L\). And we are done.

It is possible to strengthen the reverse direction of the lemma by simply stating that \(\{ f(x_n) \}\) converges without requiring a specific limit. See .

\(\displaystyle \lim_{x \to 0} \, \sin( \nicefrac{1}{x} )\) does not exist, but \(\displaystyle \lim_{x \to 0} \, x\sin( \nicefrac{1}{x} ) = 0\). See .

Proof: Let us work with \(\sin(\nicefrac{1}{x})\) first. Let us define the sequence \(x_n := \frac{1}{\pi n + \nicefrac{\pi}{2}}\). It is not hard to see that \(\lim\, x_n = 0\). Furthermore, \[\sin ( \nicefrac{1}{x_n} ) = \sin (\pi n + \nicefrac{\pi}{2}) = {(-1)}^n .\] Therefore, \(\{ \sin ( \nicefrac{1}{x_n} ) \}\) does not converge. Thus, by , \(\lim_{x \to 0} \, \sin( \nicefrac{1}{x} )\) does not exist.

Now let us look at \(x\sin(\nicefrac{1}{x})\). Let \(x_n\) be a sequence such that \(x_n \not= 0\) for all \(n\) and such that \(\lim\, x_n = 0\). Notice that \(\left\lvert {\sin(t)} \right\rvert \leq 1\) for any \(t \in {\mathbb{R}}\). Therefore, \[\left\lvert {x_n\sin(\nicefrac{1}{x_n})-0} \right\rvert = \left\lvert {x_n} \right\rvert\left\lvert {\sin(\nicefrac{1}{x_n})} \right\rvert \leq \left\lvert {x_n} \right\rvert .\] As \(x_n\) goes to 0, then \(\left\lvert {x_n} \right\rvert\) goes to zero, and hence \(\{ x_n\sin(\nicefrac{1}{x_n}) \}\) converges to zero. By , \(\displaystyle \lim_{x \to 0} \, x\sin( \nicefrac{1}{x} ) = 0\).

Keep in mind the phrase “for every sequence” in the lemma. For example, take \(\sin(\nicefrac{1}{x})\) and the sequence \(x_n = \nicefrac{1}{\pi n}\). Then \(\{ \sin (\nicefrac{1}{x_n}) \}\) is the constant zero sequence, and therefore converges to zero.

Using , we can start applying everything we know about sequential limits to limits of functions. Let us give a few important examples.

Let \(S \subset {\mathbb{R}}\) and \(c\) be a cluster point of \(S\). Let \(f \colon S \to {\mathbb{R}}\) and \(g \colon S \to {\mathbb{R}}\) be functions. Suppose the limits of \(f(x)\) and \(g(x)\) as \(x\) goes to \(c\) both exist, and that \[f(x) \leq g(x) \qquad \text{for all $x \in S$}.\] Then \[\lim_{x\to c} f(x) \leq \lim_{x\to c} g(x) .\]

Take \(\{ x_n \}\) be a sequence of numbers in \(S \setminus \{ c \}\) that converges to \(c\). Let \[L_1 := \lim_{x\to c} f(x), \qquad \text{and} \qquad L_2 := \lim_{x\to c} g(x) .\] By we know \(\{ f(x_n) \}\) converges to \(L_1\) and \(\{ g(x_n) \}\) converges to \(L_2\). We also have \(f(x_n) \leq g(x_n)\). We obtain \(L_1 \leq L_2\) using .

By applying constant functions, we get the following corollary. The proof is left as an exercise.

[fconstineq:cor] Let \(S \subset {\mathbb{R}}\) and \(c\) be a cluster point of \(S\). Let \(f \colon S \to {\mathbb{R}}\) be a function. And suppose the limit of \(f(x)\) as \(x\) goes to \(c\) exists. Suppose there are two real numbers \(a\) and \(b\) such that \[a \leq f(x) \leq b \qquad \text{for all $x \in S$}.\] Then \[a \leq \lim_{x\to c} f(x) \leq b .\]

Using in the same way as above we also get the following corollaries, whose proofs are again left as an exercise.

[fsqueeze:cor] Let \(S \subset {\mathbb{R}}\) and \(c\) be a cluster point of \(S\). Let \(f \colon S \to {\mathbb{R}}\), \(g \colon S \to {\mathbb{R}}\), and \(h \colon S \to {\mathbb{R}}\) be functions. Suppose \[f(x) \leq g(x) \leq h(x) \qquad \text{for all $x \in S$},\] and the limits of \(f(x)\) and \(h(x)\) as \(x\) goes to \(c\) both exist, and \[\lim_{x\to c} f(x) = \lim_{x\to c} h(x) .\] Then the limit of \(g(x)\) as \(x\) goes to \(c\) exists and \[\lim_{x\to c} g(x) = \lim_{x\to c} f(x) = \lim_{x\to c} h(x) .\]

[falg:cor] Let \(S \subset {\mathbb{R}}\) and \(c\) be a cluster point of \(S\). Let \(f \colon S \to {\mathbb{R}}\) and \(g \colon S \to {\mathbb{R}}\) be functions. Suppose limits of \(f(x)\) and \(g(x)\) as \(x\) goes to \(c\) both exist. Then

- \(\displaystyle \lim_{x\to c} \bigl(f(x)+g(x)\bigr) = \left(\lim_{x\to c} f(x)\right) + \left(\lim_{x\to c} g(x)\right) .\)
- \(\displaystyle \lim_{x\to c} \bigl(f(x)-g(x)\bigr) = \left(\lim_{x\to c} f(x)\right) - \left(\lim_{x\to c} g(x)\right) .\)
- \(\displaystyle \lim_{x\to c} \bigl(f(x)g(x)\bigr) = \left(\lim_{x\to c} f(x)\right) \left(\lim_{x\to c} g(x)\right) .\)
- [falg:cor:iv] If \(\displaystyle \lim_{x\to c} g(x) \not= 0\), and \(g(x) \not= 0\) for all \(x \in S \setminus \{ c \}\), then \[\lim_{x\to c} \frac{f(x)}{g(x)} = \frac{\lim_{x\to c} f(x)}{\lim_{x\to c} g(x)} .\]

#### Limits of restrictions and one-sided limits

Sometimes we work with the function defined on a subset.

Let \(f \colon S \to {\mathbb{R}}\) be a function. Let \(A \subset S\). Define the function \(f|_A \colon A \to {\mathbb{R}}\) by \[f|_A (x) := f(x) \qquad \text{for $x \in A$}.\] The function \(f|_A\) is called the *restriction* of \(f\) to \(A\).

The function \(f|_A\) is simply the function \(f\) taken on a smaller domain. The following proposition is the analogue of taking a tail of a sequence.

[prop:limrest] Let \(S \subset {\mathbb{R}}\), \(c \in {\mathbb{R}}\), and let \(f \colon S \to {\mathbb{R}}\) be a function. Suppose \(A \subset S\) is such that there is some \(\alpha > 0\) such that \(A \cap (c-\alpha,c+\alpha) = S \cap (c-\alpha,c+\alpha)\).

- The point \(c\) is a cluster point of \(A\) if and only if \(c\) is a cluster point of \(S\).
- Supposing \(c\) is a cluster point of \(S\), then \(f(x) \to L\) as \(x \to c\) if and only if \(f|_A(x) \to L\) as \(x \to c\).

First, let \(c\) be a cluster point of \(A\). Since \(A \subset S\), then if \(( A \setminus \{ c\} ) \cap (c-\epsilon,c+\epsilon)\) is nonempty for every \(\epsilon > 0\), then \(( S \setminus \{ c\} ) \cap (c-\epsilon,c+\epsilon)\) is nonempty for every \(\epsilon > 0\). Thus \(c\) is a cluster point of \(S\). Second, suppose \(c\) is a cluster point of \(S\). Then for \(\epsilon > 0\) such that \(\epsilon < \alpha\) we get that \(( A \setminus \{ c\} ) \cap (c-\epsilon,c+\epsilon) = ( S \setminus \{ c\} ) \cap (c-\epsilon,c+\epsilon)\), which is nonempty. This is true for all \(\epsilon < \alpha\) and hence \(( A \setminus \{ c\} ) \cap (c-\epsilon,c+\epsilon)\) must be nonempty for all \(\epsilon > 0\). Thus \(c\) is a cluster point of \(A\).

Now suppose \(f(x) \to L\) as \(x \to c\). That is, for every \(\epsilon > 0\) there is a \(\delta > 0\) such that if \(x \in S \setminus \{ c \}\) and \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {f(x)-L} \right\rvert < \epsilon\). Because \(A \subset S\), if \(x\) is in \(A \setminus \{ c \}\), then \(x\) is in \(S \setminus \{ c \}\), and hence \(f|_A(x) \to L\) as \(x \to c\).

Finally suppose \(f|_A(x) \to L\) as \(x \to c\). Hence for every \(\epsilon > 0\) there is a \(\delta > 0\) such that if \(x \in A \setminus \{ c \}\) and \(\left\lvert {x-c} \right\rvert < \delta\), then \(\bigl\lvert f|_A(x)-L \bigr\rvert < \epsilon\). Without loss of generality assume \(\delta \leq \alpha\). If \(\left\lvert {x-c} \right\rvert < \delta\), then \(x \in S \setminus \{c \}\) if and only if \(x \in A \setminus \{c \}\). Thus \(\left\lvert {f(x)-L} \right\rvert = \bigl\lvert f|_A(x)-L \bigr\rvert < \epsilon\).

The hypothesis of the proposition is necessary. For an arbitrary restriction we generally only get implication in only one direction, see .

A common use of restriction with respect to limits are *one-sided limits*.

[defn:onesidedlimits] Let \(f \colon S \to {\mathbb{R}}\) be function and let \(c\) be a cluster point of \(S \cap (c,\infty)\). Then if the limit of the restriction of \(f\) to \(S \cap (c,\infty)\) as \(x \to c\) exists, we define \[\lim_{x \to c^+} f(x) := \lim_{x\to c} f|_{S \cap (c,\infty)}(x) .\] Similarly if \(c\) is a cluster point of \(S \cap (-\infty,c)\) and the limit of the restriction as \(x \to c\) exists, we define \[\lim_{x \to c^-} f(x) := \lim_{x\to c} f|_{S \cap (-\infty,c)}(x) .\]

The proposition above does not apply to one-sided limits. It is possible to have one-sided limits, but no limit at a point. For example, define \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) by \(f(x) := 1\) for \(x < 0\) and \(f(x) := 0\) for \(x \geq 0\). We leave it to the reader to verify that \[\lim_{x \to 0^-} f(x) = 1, \qquad \lim_{x \to 0^+} f(x) = 0, \qquad \lim_{x \to 0} f(x) \quad \text{does not exist.}\] We have the following replacement.

[prop:onesidedlimits] Let \(S \subset {\mathbb{R}}\) be a set such that \(c\) is a cluster point of both \(S \cap (-\infty,c)\) and \(S \cap (c,\infty)\), and let \(f \colon S \to {\mathbb{R}}\) be a function. Then \[\lim_{x \to c} f(x) = L \qquad \text{if and only if} \qquad \lim_{x \to c^-} f(x) = \lim_{x \to c^+} f(x) = L .\]

That is, a limit exists if both one-sided limits exist and are equal, and vice-versa. The proof is a straightforward application of the definition of limit and is left as an exercise. The key point is that \(\bigl( S \cap (-\infty,c) \bigr) \cup \bigl( S \cap (c,\infty) \bigr) = S \setminus \{ c \}\).

## Exercises

Find the limit or prove that the limit does not exist

a) \(\displaystyle \lim_{x\to c} \sqrt{x}\), for \(c \geq 0\) | b) \(\displaystyle \lim_{x\to c} x^2+x+1\), for any \(c \in {\mathbb{R}}\) | c) \(\displaystyle \lim_{x\to 0} x^2 \cos (\nicefrac{1}{x})\) | ||

d) \(\displaystyle \lim_{x\to 0}\, \sin(\nicefrac{1}{x}) \cos (\nicefrac{1}{x})\) | e) \(\displaystyle \lim_{x\to 0}\, \sin(x) \cos (\nicefrac{1}{x})\) |

Prove .

Prove .

Prove .

Let \(A \subset S\). Show that if \(c\) is a cluster point of \(A\), then \(c\) is a cluster point of \(S\). Note the difference from .

[exercise:restrictionlimitexercise] Let \(A \subset S\). Suppose \(c\) is a cluster point of \(A\) and it is also a cluster point of \(S\). Let \(f \colon S \to {\mathbb{R}}\) be a function. Show that if \(f(x) \to L\) as \(x \to c\), then \(f|_A(x) \to L\) as \(x \to c\). Note the difference from .

Find an example of a function \(f \colon [-1,1] \to {\mathbb{R}}\) such that for \(A:=[0,1]\), the restriction \(f|_A(x) \to 0\) as \(x \to 0\), but the limit of \(f(x)\) as \(x \to 0\) does not exist. Note why you cannot apply .

Find example functions \(f\) and \(g\) such that the limit of neither \(f(x)\) nor \(g(x)\) exists as \(x \to 0\), but such that the limit of \(f(x)+g(x)\) exists as \(x \to 0\).

[exercise:contlimitcomposition] Let \(c_1\) be a cluster point of \(A \subset {\mathbb{R}}\) and \(c_2\) be a cluster point of \(B \subset {\mathbb{R}}\). Suppose \(f \colon A \to B\) and \(g \colon B \to {\mathbb{R}}\) are functions such that \(f(x) \to c_2\) as \(x \to c_1\) and \(g(y) \to L\) as \(y \to c_2\). If \(c_2 \in B\) also suppose that \(g(c_2) = L\). Let \(h(x) := g\bigl(f(x)\bigr)\) and show \(h(x) \to L\) as \(x \to c_1\). Hint: note that \(f(x)\) could equal \(c_2\) for many \(x \in A\), see also .

Let \(c\) be a cluster point of \(A \subset {\mathbb{R}}\), and \(f \colon A \to {\mathbb{R}}\) be a function. Suppose for every sequence \(\{x_n\}\) in \(A\), such that \(\lim\, x_n = c\), the sequence \(\{ f(x_n) \}_{n=1}^\infty\) is Cauchy. Prove that \(\lim_{x\to c} f(x)\) exists.

[exercise:seqflimitalt] Prove the following stronger version of one direction of : Let \(S \subset {\mathbb{R}}\), \(c\) be a cluster point of \(S\), and \(f \colon S \to {\mathbb{R}}\) be a function. Suppose that for every sequence \(\{x_n\}\) in \(S \setminus \{c\}\) such that \(\lim\, x_n = c\) the sequence \(\{ f(x_n) \}\) is convergent. Then show \(f(x) \to L\) as \(x \to c\) for some \(L \in {\mathbb{R}}\).

Prove .

Suppose \(S \subset {\mathbb{R}}\) and \(c\) is a cluster point of \(S\). Suppose \(f \colon S \to {\mathbb{R}}\) is bounded. Show that there exists a sequence \(\{ x_n \}\) with \(x_n \in S \setminus \{ c \}\) and \(\lim\, x_n = c\) such that \(\{ f(x_n) \}\) converges.

[exercise:contlimitbadcomposition] Show that the hypothesis that \(g(c_2) = L\) in is necessary. That is, find \(f\) and \(g\) such that \(f(x) \to c_2\) as \(x \to c_1\) and \(g(y) \to L\) as \(y \to c_2\), but \(g\bigl(f(x)\bigr)\) does not go to \(L\) as \(x \to c_1\).

### Continuous functions

*Note: 2–2.5 lectures*

You undoubtedly heard of continuous functions in your schooling. A high-school criterion for this concept is that a function is continuous if we can draw its graph without lifting the pen from the paper. While that intuitive concept may be useful in simple situations, we require rigor. The following definition took three great mathematicians (Bolzano, Cauchy, and finally Weierstrass) to get correctly and its final form dates only to the late 1800s.

#### Definition and basic properties

Let \(S \subset {\mathbb{R}}\), \(c \in S\), and let \(f \colon S \to {\mathbb{R}}\) be a function. We say that \(f\) is *continuous at \(c\)* if for every \(\epsilon > 0\) there is a \(\delta > 0\) such that whenever \(x \in S\) and \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {f(x)-f(c)} \right\rvert < \epsilon\).

When \(f \colon S \to {\mathbb{R}}\) is continuous at all \(c \in S\), then we simply say \(f\) is a *continuous function*.

If \(f\) is continuous for all \(c \in A\), we say \(f\) is continuous on \(A \subset S\). It is left as an easy exercise to show that this implies that \(f|_A\) is continuous, although the converse does not hold.

Continuity may be the most important definition to understand in analysis, and it is not an easy one. See . Note that \(\delta\) not only depends on \(\epsilon\), but also on \(c\); we need not pick one \(\delta\) for all \(c \in S\). It is no accident that the definition of continuity is similar to the definition of a limit of a function. The main feature of continuous functions is that these are precisely the functions that behave nicely with limits.

[contbasic:prop] Suppose \(f \colon S \to {\mathbb{R}}\) is a function and \(c \in S\). Then

- If \(c\) is not a cluster point of \(S\), then \(f\) is continuous at \(c\).
- If \(c\) is a cluster point of \(S\), then \(f\) is continuous at \(c\) if and only if the limit of \(f(x)\) as \(x \to c\) exists and \[\lim_{x\to c} f(x) = f(c) .\]
- \(f\) is continuous at \(c\) if and only if for every sequence \(\{ x_n \}\) where \(x_n \in S\) and \(\lim\, x_n = c\), the sequence \(\{ f(x_n) \}\) converges to \(f(c)\).

Let us start with the first item. Suppose \(c\) is not a cluster point of \(S\). Then there exists a \(\delta > 0\) such that \(S \cap (c-\delta,c+\delta) = \{ c \}\). Therefore, for any \(\epsilon > 0\), simply pick this given delta. The only \(x \in S\) such that \(\left\lvert {x-c} \right\rvert < \delta\) is \(x=c\). Then \(\left\lvert {f(x)-f(c)} \right\rvert = \left\lvert {f(c)-f(c)} \right\rvert = 0 < \epsilon\).

Let us move to the second item. Suppose \(c\) is a cluster point of \(S\). Let us first suppose that \(\lim_{x\to c} f(x) = f(c)\). Then for every \(\epsilon > 0\) there is a \(\delta > 0\) such that if \(x \in S \setminus \{ c \}\) and \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {f(x)-f(c)} \right\rvert < \epsilon\). As \(\left\lvert {f(c)-f(c)} \right\rvert = 0 < \epsilon\), then the definition of continuity at \(c\) is satisfied. On the other hand, suppose \(f\) is continuous at \(c\). For every \(\epsilon > 0\), there exists a \(\delta > 0\) such that for \(x \in S\) where \(\left\lvert {x-c} \right\rvert < \delta\) we have \(\left\lvert {f(x)-f(c)} \right\rvert < \epsilon\). Then the statement is, of course, still true if \(x \in S \setminus \{ c \} \subset S\). Therefore \(\lim_{x\to c} f(x) = f(c)\).

For the third item, suppose \(f\) is continuous at \(c\). Let \(\{ x_n \}\) be a sequence such that \(x_n \in S\) and \(\lim\, x_n = c\). Let \(\epsilon > 0\) be given. Find a \(\delta > 0\) such that \(\left\lvert {f(x)-f(c)} \right\rvert < \epsilon\) for all \(x \in S\) where \(\left\lvert {x-c} \right\rvert < \delta\). Find an \(M \in {\mathbb{N}}\) such that for \(n \geq M\) we have \(\left\lvert {x_n-c} \right\rvert < \delta\). Then for \(n \geq M\) we have that \(\left\lvert {f(x_n)-f(c)} \right\rvert < \epsilon\), so \(\{ f(x_n) \}\) converges to \(f(c)\).

Let us prove the converse of the third item by contrapositive. Suppose \(f\) is not continuous at \(c\). Then there exists an \(\epsilon > 0\) such that for all \(\delta > 0\), there exists an \(x \in S\) such that \(\left\lvert {x-c} \right\rvert < \delta\) and \(\left\lvert {f(x)-f(c)} \right\rvert \geq \epsilon\). Let us define a sequence \(\{ x_n \}\) as follows. Let \(x_n \in S\) be such that \(\left\lvert {x_n-c} \right\rvert < \nicefrac{1}{n}\) and \(\left\lvert {f(x_n)-f(c)} \right\rvert \geq \epsilon\). Now \(\{ x_n \}\) is a sequence of numbers in \(S\) such that \(\lim\, x_n = c\) and such that \(\left\lvert {f(x_n)-f(c)} \right\rvert \geq \epsilon\) for all \(n \in {\mathbb{N}}\). Thus \(\{ f(x_n) \}\) does not converge to \(f(c)\). It may or may not converge, but it definitely does not converge to \(f(c)\).

The last item in the proposition is particularly powerful. It allows us to quickly apply what we know about limits of sequences to continuous functions and even to prove that certain functions are continuous. It can also be strengthened, see .

\(f \colon (0,\infty) \to {\mathbb{R}}\) defined by \(f(x) := \nicefrac{1}{x}\) is continuous.

Proof: Fix \(c \in (0,\infty)\). Let \(\{ x_n \}\) be a sequence in \((0,\infty)\) such that \(\lim\, x_n = c\). Then we know that \[f(c) = \frac{1}{c} = \frac{1}{\lim\, x_n} = \lim_{n \to \infty} \frac{1}{x_n} = \lim_{n \to \infty} f(x_n) .\] Thus \(f\) is continuous at \(c\). As \(f\) is continuous at all \(c \in (0,\infty)\), \(f\) is continuous.

We have previously shown \(\lim_{x \to c} x^2 = c^2\) directly. Therefore the function \(x^2\) is continuous. We can use the continuity of algebraic operations with respect to limits of sequences, which we proved in the previous chapter, to prove a much more general result.

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be a *polynomial*. That is \[f(x) = a_d x^d + a_{d-1} x^{d-1} + \cdots + a_1 x + a_0 ,\] for some constants \(a_0, a_1, \ldots, a_d\). Then \(f\) is continuous.

Fix \(c \in {\mathbb{R}}\). Let \(\{ x_n \}\) be a sequence such that \(\lim\, x_n = c\). Then \[\begin{split} f(c) &= a_d c^d + a_{d-1} c^{d-1} + \cdots + a_1 c + a_0 \\ &= a_d {(\lim\, x_n)}^d + a_{d-1} {(\lim\, x_n)}^{d-1} + \cdots + a_1 (\lim\, x_n) + a_0 \\ & = \lim_{n \to \infty} \left( a_d x_n^d + a_{d-1} x_n^{d-1} + \cdots + a_1 x_n + a_0 \right) = \lim_{n \to \infty} f(x_n) . \end{split}\] Thus \(f\) is continuous at \(c\). As \(f\) is continuous at all \(c \in {\mathbb{R}}\), \(f\) is continuous.

By similar reasoning, or by appealing to , we can prove the following. The details of the proof are left as an exercise.

[contalg:prop] Let \(f \colon S \to {\mathbb{R}}\) and \(g \colon S \to {\mathbb{R}}\) be functions continuous at \(c \in S\).

- The function \(h \colon S \to {\mathbb{R}}\) defined by \(h(x) := f(x)+g(x)\) is continuous at \(c\).
- The function \(h \colon S \to {\mathbb{R}}\) defined by \(h(x) := f(x)-g(x)\) is continuous at \(c\).
- The function \(h \colon S \to {\mathbb{R}}\) defined by \(h(x) := f(x)g(x)\) is continuous at \(c\).
- If \(g(x)\not=0\) for all \(x \in S\), the function \(h \colon S \to {\mathbb{R}}\) defined by \(h(x) := \frac{f(x)}{g(x)}\) is continuous at \(c\).

[sincos:example] The functions \(\sin(x)\) and \(\cos(x)\) are continuous. In the following computations we use the sum-to-product trigonometric identities. We also use the simple facts that \(\left\lvert {\sin(x)} \right\rvert \leq \left\lvert {x} \right\rvert\), \(\left\lvert {\cos(x)} \right\rvert \leq 1\), and \(\left\lvert {\sin(x)} \right\rvert \leq 1\). \[\begin{split} \left\lvert {\sin(x)-\sin(c)} \right\rvert & = \left\lvert { 2 \sin \left( \frac{x-c}{2} \right) \cos \left( \frac{x+c}{2} \right) } \right\rvert \\ & = 2 \left\lvert { \sin \left( \frac{x-c}{2} \right) } \right\rvert \left\lvert { \cos \left( \frac{x+c}{2} \right) } \right\rvert \\ & \leq 2 \left\lvert { \sin \left( \frac{x-c}{2} \right) } \right\rvert \\ & \leq 2 \left\lvert { \frac{x-c}{2} } \right\rvert = \left\lvert {x-c} \right\rvert \end{split}\] \[\begin{split} \left\lvert {\cos(x)-\cos(c)} \right\rvert & = \left\lvert { -2 \sin \left( \frac{x-c}{2} \right) \sin \left( \frac{x+c}{2} \right) } \right\rvert \\ & = 2 \left\lvert { \sin \left( \frac{x-c}{2} \right) } \right\rvert \left\lvert { \sin \left( \frac{x+c}{2} \right) } \right\rvert \\ & \leq 2 \left\lvert { \sin \left( \frac{x-c}{2} \right) } \right\rvert \\ & \leq 2 \left\lvert { \frac{x-c}{2} } \right\rvert = \left\lvert {x-c} \right\rvert \end{split}\]

The claim that sin and cos are continuous follows by taking an arbitrary sequence \(\{ x_n \}\) converging to \(c\), or by applying the definition of continuity directly. Details are left to the reader.

#### Composition of continuous functions

You have probably already realized that one of the basic tools in constructing complicated functions out of simple ones is composition. A useful property of continuous functions is that compositions of continuous functions are again continuous. Recall that for two functions \(f\) and \(g\), the composition \(f \circ g\) is defined by \((f \circ g)(x) := f\bigl(g(x)\bigr)\).

Let \(A, B \subset {\mathbb{R}}\) and \(f \colon B \to {\mathbb{R}}\) and \(g \colon A \to B\) be functions. If \(g\) is continuous at \(c \in A\) and \(f\) is continuous at \(g(c)\), then \(f \circ g \colon A \to {\mathbb{R}}\) is continuous at \(c\).

Let \(\{ x_n \}\) be a sequence in \(A\) such that \(\lim\, x_n = c\). Then as \(g\) is continuous at \(c\), then \(\{ g(x_n) \}\) converges to \(g(c)\). As \(f\) is continuous at \(g(c)\), then \(\{ f\bigl(g(x_n)\bigr) \}\) converges to \(f\bigl(g(c)\bigr)\). Thus \(f \circ g\) is continuous at \(c\).

Claim: \({\bigl(\sin(\nicefrac{1}{x})\bigr)}^2\) is a continuous function on \((0,\infty)\).

Proof: First note that \(\nicefrac{1}{x}\) is a continuous function on \((0,\infty)\) and \(\sin(x)\) is a continuous function on \((0,\infty)\) (actually on all of \({\mathbb{R}}\), but \((0,\infty)\) is the range for \(\nicefrac{1}{x}\)). Hence the composition \(\sin(\nicefrac{1}{x})\) is continuous. We also know that \(x^2\) is continuous on the interval \((-1,1)\) (the range of sin). Thus the composition \({\bigl(\sin(\nicefrac{1}{x})\bigr)}^2\) is also continuous on \((0,\infty)\).

#### Discontinuous functions

When \(f\) is not continuous at \(c\), we say \(f\) is *discontinuous* at \(c\), or that it has a *discontinuity* at \(c\). If we state the contrapositive of the third item of as a separate claim we get an easy to use test for discontinuities.

Let \(f \colon S \to {\mathbb{R}}\) be a function. Suppose that for some \(c \in S\), there exists a sequence \(\{ x_n \}\), \(x_n \in S\), and \(\lim\, x_n = c\) such that \(\{ f(x_n) \}\) does not converge to \(f(c)\) (or does not converge at all), then \(f\) is not continuous at \(c\).

[example:stepdiscont] The function \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) defined by \[f(x) := \begin{cases} -1 & \text{ if $x < 0$,} \\ 1 & \text{ if $x \geq 0$,} \end{cases}\] is not continuous at 0.

Proof: Take the sequence \(\{ - \nicefrac{1}{n} \}\). Then \(f(-\nicefrac{1}{n}) = -1\) and so \(\lim\, f(-\nicefrac{1}{n}) = -1\), but \(f(0) = 1\).

For an extreme example we take the so-called *Dirichlet function*. \[f(x) := \begin{cases} 1 & \text{ if $x$ is rational,} \\ 0 & \text{ if $x$ is irrational.} \end{cases}\] The function \(f\) is discontinuous at all \(c \in {\mathbb{R}}\).

Proof: Suppose \(c\) is rational. Take a sequence \(\{ x_n \}\) of irrational numbers such that \(\lim\, x_n = c\) (why can we?). Then \(f(x_n) = 0\) and so \(\lim\, f(x_n) = 0\), but \(f(c) = 1\). If \(c\) is irrational, take a sequence of rational numbers \(\{ x_n \}\) that converges to \(c\) (why can we?). Then \(\lim\, f(x_n) = 1\), but \(f(c) = 0\).

Let us yet again test the limits of your intuition. Can there exist a function that is continuous on all irrational numbers, but discontinuous at all rational numbers? There are rational numbers arbitrarily close to any irrational number. Perhaps strangely, the answer is yes. The following example is called the *Thomae function*^{15} or the *popcorn function*.

[popcornfunction:example] Let \(f \colon (0,1) \to {\mathbb{R}}\) be defined by \[f(x) := \begin{cases} \nicefrac{1}{k} & \text{ if $x=\nicefrac{m}{k}$ where $m,k \in {\mathbb{N}}$ and $m$ and $k$ have no common divisors,} \\ 0 & \text{ if $x$ is irrational}. \end{cases}\] Then \(f\) is continuous at all irrational \(c \in (0,1)\) and discontinuous at all rational \(c\). See the graph of \(f\) in .

Proof: Suppose \(c = \nicefrac{m}{k}\) is rational. Take a sequence of irrational numbers \(\{ x_n \}\) such that \(\lim\, x_n = c\). Then \(\lim\, f(x_n) = \lim \, 0 = 0\), but \(f(c) = \nicefrac{1}{k} \not= 0\). So \(f\) is discontinuous at \(c\).

Now let \(c\) be irrational, so \(f(c) = 0\). Take a sequence \(\{ x_n \}\) of numbers in \((0,1)\) such that \(\lim\, x_n = c\). Given \(\epsilon > 0\), find \(K \in {\mathbb{N}}\) such that \(\nicefrac{1}{K} < \epsilon\) by the . If \(\nicefrac{m}{k} \in (0,1)\) is lowest terms (no common divisors), then \(m < k\). So there are only finitely many rational numbers in \((0,1)\) whose denominator \(k\) in lowest terms is less than \(K\). Hence there is an \(M\) such that for \(n \geq M\), all the numbers \(x_n\) that are rational have a denominator larger than or equal to \(K\). Thus for \(n \geq M\) \[\left\lvert {f(x_n) - 0} \right\rvert = f(x_n) \leq \nicefrac{1}{K} < \epsilon .\] Therefore \(f\) is continuous at irrational \(c\).

Let us end on an easier example.

Define \(g \colon {\mathbb{R}}\to {\mathbb{R}}\) by \(g(x) := 0\) if \(x \not= 0\) and \(g(0) := 1\). Then \(g\) is not continuous at zero, but continuous everywhere else (why?). The point \(x=0\) is called a *removable discontinuity*. That is because if we would change the definition of \(g\), by insisting that \(g(0)\) be \(0\), we would obtain a continuous function. On the other hand let \(f\) be the function of example . Then \(f\) does not have a removable discontinuity at \(0\). No matter how we would define \(f(0)\) the function will still fail to be continuous. The difference is that \(\lim_{x\to 0} g(x)\) exists while \(\lim_{x\to 0} f(x)\) does not.

Let us stay with this example but show another phenomenon. Let \(A = \{ 0 \}\), then \(g|_A\) is continuous (why?), while \(g\) is not continuous on \(A\).

## Exercises

Using the definition of continuity directly prove that \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) defined by \(f(x) := x^2\) is continuous.

Using the definition of continuity directly prove that \(f \colon (0,\infty) \to {\mathbb{R}}\) defined by \(f(x) := \nicefrac{1}{x}\) is continuous.

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be defined by \[f(x) := \begin{cases} x & \text{ if $x$ is rational,} \\ x^2 & \text{ if $x$ is irrational.} \end{cases}\] Using the definition of continuity directly prove that \(f\) is continuous at \(1\) and discontinuous at \(2\).

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be defined by \[f(x) := \begin{cases} \sin(\nicefrac{1}{x}) & \text{ if $x \not= 0$,} \\ 0 & \text{ if $x=0$.} \end{cases}\] Is \(f\) continuous? Prove your assertion.

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be defined by \[f(x) := \begin{cases} x \sin(\nicefrac{1}{x}) & \text{ if $x \not= 0$,} \\ 0 & \text{ if $x=0$.} \end{cases}\] Is \(f\) continuous? Prove your assertion.

Prove .

Prove the following statement. Let \(S \subset {\mathbb{R}}\) and \(A \subset S\). Let \(f \colon S \to {\mathbb{R}}\) be a continuous function. Then the restriction \(f|_A\) is continuous.

Suppose \(S \subset {\mathbb{R}}\). Suppose for some \(c \in {\mathbb{R}}\) and \(\alpha > 0\), we have \(A=(c-\alpha,c+\alpha) \subset S\). Let \(f \colon S \to {\mathbb{R}}\) be a function. Prove that if \(f|_A\) is continuous at \(c\), then \(f\) is continuous at \(c\).

Give an example of functions \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) and \(g \colon {\mathbb{R}}\to {\mathbb{R}}\) such that the function \(h\) defined by \(h(x) := f(x) + g(x)\) is continuous, but \(f\) and \(g\) are not continuous. Can you find \(f\) and \(g\) that are nowhere continuous, but \(h\) is a continuous function?

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) and \(g \colon {\mathbb{R}}\to {\mathbb{R}}\) be continuous functions. Suppose that for all rational numbers \(r\), \(f(r) = g(r)\). Show that \(f(x) = g(x)\) for all \(x\).

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be continuous. Suppose \(f(c) > 0\). Show that there exists an \(\alpha > 0\) such that for all \(x \in (c-\alpha,c+\alpha)\) we have \(f(x) > 0\).

Let \(f \colon {\mathbb{Z}}\to {\mathbb{R}}\) be a function. Show that \(f\) is continuous.

[exercise:contseqalt] Let \(f \colon S \to {\mathbb{R}}\) be a function and \(c \in S\), such that for every sequence \(\{ x_n \}\) in \(S\) with \(\lim\, x_n = c\), the sequence \(\{ f(x_n) \}\) converges. Show that \(f\) is continuous at \(c\).

Suppose \(f \colon [-1,0] \to {\mathbb{R}}\) and \(g \colon [0,1] \to {\mathbb{R}}\) are continuous and \(f(0) = g(0)\). Define \(h \colon [-1,1] \to {\mathbb{R}}\) by \(h(x) := f(x)\) if \(x \leq 0\) and \(h(x) := g(x)\) if \(x > 0\). Show that \(h\) is continuous.

Suppose \(g \colon {\mathbb{R}}\to {\mathbb{R}}\) is a continuous function such that \(g(0) = 0\), and supppse \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is such that \(\left\lvert {f(x)-f(y)} \right\rvert \leq g(x-y)\) for all \(x\) and \(y\). Show that \(f\) is continuous.

Suppose \(f(x+y) = f(x) + f(y)\) for some \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) such that \(f\) is continuous at 0. Show that \(f(x) = ax\) for some \(a \in {\mathbb{R}}\). Hint: Show that \(f(nx) = nf(x)\), then show \(f\) is continuous on \({\mathbb{R}}\). Then show that \(\nicefrac{f(x)}{x} = f(1)\) for all rational \(x\).

### Min-max and intermediate value theorems

*Note: 1.5 lectures*

Continuous functions defined on closed and bounded intervals have some interesting and very useful properties.

#### Min-max theorem

Recall a function \(f \colon [a,b] \to {\mathbb{R}}\) is *bounded* if there exists a \(B \in {\mathbb{R}}\) such that \(\left\lvert {f(x)} \right\rvert \leq B\) for all \(x \in [a,b]\). We have the following lemma.

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Then \(f\) is bounded.

Let us prove this claim by contrapositive. Suppose \(f\) is not bounded, then for each \(n \in {\mathbb{N}}\), there is an \(x_n \in [a,b]\), such that \[\left\lvert {f(x_n)} \right\rvert \geq n .\] Now \(\{ x_n \}\) is a bounded sequence as \(a \leq x_n \leq b\). By the , there is a convergent subsequence \(\{ x_{n_i} \}\). Let \(x := \lim\, x_{n_i}\). Since \(a \leq x_{n_i} \leq b\) for all \(i\), then \(a \leq x \leq b\). The limit \(\lim\, f(x_{n_i})\) does not exist as the sequence is not bounded as \(\left\lvert {f(x_{n_i})} \right\rvert \geq n_i \geq i\). On the other hand \(f(x)\) is a finite number and \[f(x) = f\left( \lim_{i\to\infty} x_{n_i} \right) .\] Thus \(f\) is not continuous at \(x\).

In fact, for a continuous \(f\), we will see that the minimum and the maximum are actually achieved. Recall from calculus that \(f \colon S \to {\mathbb{R}}\) achieves an *absolute minimum* at \(c \in S\) if \[f(x) \geq f(c) \qquad \text{ for all $x \in S$.}\] On the other hand, \(f\) achieves an *absolute maximum* at \(c \in S\) if \[f(x) \leq f(c) \qquad \text{ for all $x \in S$.}\] We say \(f\) achieves an absolute minimum or an absolute maximum on \(S\) if such a \(c \in S\) exists. If \(S\) is a closed and bounded interval, then a continuous \(f\) must have an absolute minimum and an absolute maximum on \(S\).

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Then \(f\) achieves both an absolute minimum and an absolute maximum on \([a,b]\).

We have shown that \(f\) is bounded by the lemma. Therefore, the set \(f([a,b]) = \{ f(x) : x \in [a,b] \}\) has a supremum and an infimum. From what we know about suprema and infima, there exist sequences in the set \(f([a,b])\) that approach them. That is, there are sequences \(\{ f(x_n) \}\) and \(\{ f(y_n) \}\), where \(x_n, y_n\) are in \([a,b]\), such that \[\lim_{n\to\infty} f(x_n) = \inf f([a,b]) \qquad \text{and} \qquad \lim_{n\to\infty} f(y_n) = \sup f([a,b]).\] We are not done yet, we need to find where the minimum and the maxima are. The problem is that the sequences \(\{ x_n \}\) and \(\{ y_n \}\) need not converge. We know \(\{ x_n \}\) and \(\{ y_n \}\) are bounded (their elements belong to a bounded interval \([a,b]\)). We apply the . Hence there exist convergent subsequences \(\{ x_{n_i} \}\) and \(\{ y_{m_i} \}\). Let \[x := \lim_{i\to\infty} x_{n_i} \qquad \text{and} \qquad y := \lim_{i\to\infty} y_{m_i}.\] Then as \(a \leq x_{n_i} \leq b\), we have that \(a \leq x \leq b\). Similarly \(a \leq y \leq b\), so \(x\) and \(y\) are in \([a,b]\). We apply that a limit of a subsequence is the same as the limit of the sequence, and we apply the continuity of \(f\) to obtain \[\inf f([a,b]) = \lim_{n\to\infty} f(x_n) = \lim_{i\to\infty} f(x_{n_i}) = f \left( \lim_{i\to\infty} x_{n_i} \right) = f(x) .\] Similarly, \[\sup f([a,b]) = \lim_{n\to\infty} f(m_n) = \lim_{i\to\infty} f(y_{m_i}) = f \left( \lim_{i\to\infty} y_{m_i} \right) = f(y) .\] Therefore, \(f\) achieves an absolute minimum at \(x\) and \(f\) achieves an absolute maximum at \(y\).

The function \(f(x) := x^2+1\) defined on the interval \([-1,2]\) achieves a minimum at \(x=0\) when \(f(0) = 1\). It achieves a maximum at \(x=2\) where \(f(2) = 5\). Do note that the domain of definition matters. If we instead took the domain to be \([-10,10]\), then \(x=2\) would no longer be a maximum of \(f\). Instead the maximum would be achieved at either \(x=10\) or \(x=-10\).

Let us show by examples that the different hypotheses of the theorem are truly needed.

The function \(f(x) := x\), defined on the whole real line, achieves neither a minimum, nor a maximum. So it is important that we are looking at a bounded interval.

The function \(f(x) := \nicefrac{1}{x}\), defined on \((0,1)\) achieves neither a minimum, nor a maximum. The values of the function are unbounded as we approach 0. Also as we approach \(x=1\), the values of the function approach 1, but \(f(x) > 1\) for all \(x \in (0,1)\). There is no \(x \in (0,1)\) such that \(f(x) = 1\). So it is important that we are looking at a closed interval.

Continuity is important. Define \(f \colon [0,1] \to {\mathbb{R}}\) by \(f(x) := \nicefrac{1}{x}\) for \(x > 0\) and let \(f(0) := 0\). Then the function does not achieve a maximum. The problem is that the function is not continuous at 0.

#### Bolzano’s intermediate value theorem

Bolzano’s intermediate value theorem is one of the cornerstones of analysis. It is sometimes called only intermediate value theorem, or just Bolzano’s theorem. To prove Bolzano’s theorem we prove the following simpler lemma.

[IVT:lemma] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Suppose \(f(a) < 0\) and \(f(b) > 0\). Then there exists a number \(c \in (a,b)\) such that \(f(c) = 0\).

We define two sequences \(\{ a_n \}\) and \(\{ b_n \}\) inductively:

- Let \(a_1 := a\) and \(b_1 := b\).
- If \(f\left(\frac{a_n+b_n}{2}\right) \geq 0\), let \(a_{n+1} := a_n\) and \(b_{n+1} := \frac{a_n+b_n}{2}\).
- If \(f\left(\frac{a_n+b_n}{2}\right) < 0\), let \(a_{n+1} := \frac{a_n+b_n}{2}\) and \(b_{n+1} := b_n\).

See for an example defining the first five steps. From the definition of the two sequences it is obvious that if \(a_n < b_n\), then \(a_{n+1} < b_{n+1}\). Thus by \(a_n < b_n\) for all \(n\). Furthermore, \(a_n \leq a_{n+1}\) and \(b_n \geq b_{n+1}\) for all \(n\), that is the sequences are monotone. As \(a_n < b_n \leq b_1 = b\) and \(b_n > a_n \geq a_1 = a\) for all \(n\), the sequences are also bounded. Therefore, the sequences converge. Let \(c := \lim\, a_n\) and \(d := \lim\, b_n\). We now want to show that \(c=d\). We notice \[b_{n+1} - a_{n+1} = \frac{b_n-a_n}{2}.\] By we see that \[b_n - a_n = \frac{b_1-a_1}{2^{n-1}} = 2^{1-n} (b-a) .\] As \(2^{1-n}(b-a)\) converges to zero, we take the limit as \(n\) goes to infinity to get \[d-c = \lim_{n\to\infty} (b_n - a_n) = \lim_{n\to\infty} 2^{1-n} (b-a) = 0.\] In other words \(d=c\).

By construction, for all \(n\) we have \[f(a_n) < 0 \qquad \text{and} \qquad f(b_n) \geq 0 .\] We use the fact that \(\lim\, a_n = \lim\, b_n = c\) and the continuity of \(f\) to take limits in those inequalities to get \[f(c) = \lim\, f(a_n) \leq 0 \qquad \text{and} \qquad f(c) = \lim\, f(b_n) \geq 0 .\] As \(f(c) \geq 0\) and \(f(c) \leq 0\), we conclude \(f(c) = 0\). Obviously, \(a < c < b\).

Notice that the proof tells us how to find the \(c\). The proof is not only useful for us pure mathematicians, but it is a useful idea in applied mathematics.

[IVT:thm] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Suppose there exists a \(y\) such that \(f(a) < y < f(b)\) or \(f(a) > y > f(b)\). Then there exists a \(c \in (a,b)\) such that \(f(c) = y\).

The theorem says that a continuous function on a closed interval achieves all the values between the values at the endpoints.

If \(f(a) < y < f(b)\), then define \(g(x) := f(x)-y\). Then we see that \(g(a) < 0\) and \(g(b) > 0\) and we can apply to \(g\). If \(g(c) = 0\), then \(f(c) = y\).

Similarly if \(f(a) > y > f(b)\), then define \(g(x) := y-f(x)\). Then again \(g(a) < 0\) and \(g(b) > 0\) and we can apply . Again if \(g(c) = 0\), then \(f(c) = y\).

If a function is continuous, then the restriction to a subset is continuous. So if \(f \colon S \to {\mathbb{R}}\) is continuous and \([a,b] \subset S\), then \(f|_{[a,b]}\) is also continuous. Hence, we generally apply the theorem to a function continuous on some large set \(S\), but we restrict attention to an interval.

The polynomial \(f(x) := x^3-2x^2+x-1\) has a real root in \((1,2)\). We simply notice that \(f(1) = -1\) and \(f(2) = 1\). Hence there must exist a point \(c \in (1,2)\) such that \(f(c) = 0\). To find a better approximation of the root we could follow the proof of . For example, next we would look at 1.5 and find that \(f(1.5) = -0.625\). Therefore, there is a root of the equation in \((1.5,2)\). Next we look at 1.75 and note that \(f(1.75) \approx -0.016\). Hence there is a root of \(f\) in \((1.75,2)\). Next we look at 1.875 and find that \(f(1.875) \approx 0.44\), thus there is a root in \((1.75,1.875)\). We follow this procedure until we gain sufficient precision.

The technique above is the simplest method of finding roots of polynomials. Finding roots of polynomials is perhaps the most common problem in applied mathematics. In general it is hard to do quickly, precisely and automatically. We can use the intermediate value theorem to find roots for any continuous function, not just a polynomial.

There are better and faster methods of finding roots of equations, such as Newton’s method. One advantage of the above method is its simplicity. The moment we find an initial interval where the intermediate value theorem applies, we are guaranteed to find a root up to a desired precision in finitely many steps. Furthermore, the method only requires a continuous function.

The theorem guarantees at least one \(c\) such that \(f(c) = y\), but there may be many different roots of the equation \(f(c) = y\). If we follow the procedure of the proof, we are guaranteed to find approximations to one such root. We need to work harder to find any other roots.

Polynomials of even degree may not have any real roots. For example, there is no real number \(x\) such that \(x^2+1 = 0\). Odd polynomials, on the other hand, always have at least one real root.

Let \(f(x)\) be a polynomial of odd degree. Then \(f\) has a real root.

Suppose \(f\) is a polynomial of odd degree \(d\). We write \[f(x) = a_d x^d + a_{d-1} x^{d-1} + \cdots + a_1 x + a_0 ,\] where \(a_d \not= 0\). We divide by \(a_d\) to obtain a polynomial \[g(x) := x^d + b_{d-1} x^{d-1} + \cdots + b_1 x + b_0 ,\] where \(b_k = \nicefrac{a_k}{a_d}\). Let us show that \(g(n)\) is positive for some large \(n \in {\mathbb{N}}\). \[\begin{split} \left\lvert {\frac{b_{d-1} n^{d-1} + \cdots + b_1 n + b_0}{n^d}} \right\rvert & = \frac{\left\lvert {b_{d-1} n^{d-1} + \cdots + b_1 n + b_0} \right\rvert}{n^d} \\ & \leq \frac{\left\lvert {b_{d-1}} \right\rvert n^{d-1} + \cdots + \left\lvert {b_1} \right\rvert n + \left\lvert {b_0} \right\rvert}{n^d} \\ & \leq \frac{\left\lvert {b_{d-1}} \right\rvert n^{d-1} + \cdots + \left\lvert {b_1} \right\rvert n^{d-1} + \left\lvert {b_0} \right\rvert n^{d-1}}{n^d} \\ & = \frac{n^{d-1}\bigl(\left\lvert {b_{d-1}} \right\rvert + \cdots + \left\lvert {b_1} \right\rvert + \left\lvert {b_0} \right\rvert\bigr)}{n^d} \\ & = \frac{1}{n} \bigl(\left\lvert {b_{d-1}} \right\rvert + \cdots + \left\lvert {b_1} \right\rvert + \left\lvert {b_0} \right\rvert\bigr) . \end{split}\] Therefore \[\lim_{n\to\infty} \frac{b_{d-1} n^{d-1} + \cdots + b_1 n + b_0}{n^d} = 0 .\] Thus there exists an \(M \in {\mathbb{N}}\) such that \[\left\lvert {\frac{b_{d-1} M^{d-1} + \cdots + b_1 M + b_0}{M^d}} \right\rvert < 1 ,\] which implies \[-(b_{d-1} M^{d-1} + \cdots + b_1 M + b_0) < M^d .\] Therefore \(g(M) > 0\).

Next we look at \(g(-n)\) for \(n \in {\mathbb{N}}\). By a similar argument (exercise) we find that there exists some \(K \in {\mathbb{N}}\) such that \(b_{d-1} {(-K)}^{d-1} + \cdots + b_1 (-K) + b_0 < K^d\) and therefore \(g(-K) < 0\) (why?). In the proof make sure you use the fact that \(d\) is odd. In particular, if \(d\) is odd then \({(-n)}^d = -(n^d)\).

We appeal to the intermediate value theorem, to find a \(c \in [-K,M]\) such that \(g(c) = 0\). As \(g(x) = \frac{f(x)}{a_d}\), we see that \(f(c) = 0\), and the proof is done.

An interesting fact is that there do exist discontinuous functions that have the intermediate value property. The function \[f(x) := \begin{cases} \sin(\nicefrac{1}{x}) & \text{ if $x \not= 0$,} \\ 0 & \text{ if $x=0$,} \end{cases}\] is not continuous at 0, however, it has the intermediate value property. That is, for any \(a < b\), and any \(y\) such that \(f(a) < y < f(b)\) or \(f(a) > y > f(b)\), there exists a \(c\) such that \(f(y) = c\). Proof is left as an exercise.

## Exercises

Find an example of a discontinuous function \(f \colon [0,1] \to {\mathbb{R}}\) where the intermediate value theorem fails.

Find an example of a *bounded* discontinuous function \(f \colon [0,1] \to {\mathbb{R}}\) that has neither an absolute minimum nor an absolute maximum.

Let \(f \colon (0,1) \to {\mathbb{R}}\) be a continuous function such that \(\displaystyle \lim_{x\to 0} f(x) = \displaystyle \lim_{x\to 1} f(x) = 0\). Show that \(f\) achieves either an absolute minimum or an absolute maximum on \((0,1)\) (but perhaps not both).

Let \[f(x) := \begin{cases} \sin(\nicefrac{1}{x}) & \text{ if $x \not= 0$,} \\ 0 & \text{ if $x=0$.} \end{cases}\] Show that \(f\) has the intermediate value property. That is, for any \(a < b\), if there exists a \(y\) such that \(f(a) < y < f(b)\) or \(f(a) > y > f(b)\), then there exists a \(c \in (a,b)\) such that \(f(c) = y\).

Suppose \(g(x)\) is a polynomial of odd degree \(d\) such that \[g(x) = x^d + b_{d-1} x^{d-1} + \cdots + b_1 x + b_0 ,\] for some real numbers \(b_{0}, b_1, \ldots, b_{d-1}\). Show that there exists a \(K \in {\mathbb{N}}\) such that \(g(-K) < 0\). Hint: Make sure to use the fact that \(d\) is odd. You will have to use that \({(-n)}^d = -(n^d)\).

Suppose \(g(x)\) is a polynomial of positive even degree \(d\) such that \[g(x) = x^d + b_{d-1} x^{d-1} + \cdots + b_1 x + b_0 ,\] for some real numbers \(b_{0}, b_1, \ldots, b_{d-1}\). Suppose \(g(0) < 0\). Show that \(g\) has at least two distinct real roots.

[exercise:imageofinterval] Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is a continuous function. Prove that the direct image \(f([a,b])\) is a closed and bounded interval or a single number.

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is continuous and periodic with period \(P > 0\). That is, \(f(x+P) = f(x)\) for all \(x \in {\mathbb{R}}\). Show that \(f\) achieves an absolute minimum and an absolute maximum.

Suppose \(f(x)\) is a bounded polynomial, in other words, there is an \(M\) such that \(\left\lvert {f(x)} \right\rvert \leq M\) for all \(x \in {\mathbb{R}}\). Prove that \(f\) must be a constant.

Suppose \(f \colon [0,1] \to [0,1]\) is continuous. Show that \(f\) has a fixed point, in other words, show that there exists an \(x \in [0,1]\) such that \(f(x) = x\).

Find an example of a bounded function \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) that does not achieve an absolute minimum nor an absolute maximum on \({\mathbb{R}}\).

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a continuous function such that \(x \leq f(x) \leq x+1\) for all \(x \in {\mathbb{R}}\). Find \(f({\mathbb{R}})\).

True/False, prove or find a counterexample. If \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a continuous function such that \(f|_

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[19]/p[13]/span, line 1, column 1 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

### Uniform continuity

*Note: 1.5–2 lectures (Continuous extension and Lipschitz can be optional)*

#### Uniform continuity

We made a fuss of saying that the \(\delta\) in the definition of continuity depended on the point \(c\). There are situations when it is advantageous to have a \(\delta\) independent of any point. Let us give a name to this concept.

Let \(S \subset {\mathbb{R}}\), and let \(f \colon S \to {\mathbb{R}}\) be a function. Suppose for any \(\epsilon > 0\) there exists a \(\delta > 0\) such that whenever \(x, c \in S\) and \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {f(x)-f(c)} \right\rvert < \epsilon\). Then we say \(f\) is *uniformly continuous*.

It is not hard to see that a uniformly continuous function must be continuous. The only difference in the definitions is that for a given \(\epsilon > 0\) we pick a \(\delta > 0\) that works for all \(c \in S\). That is, \(\delta\) can no longer depend on \(c\), it only depends on \(\epsilon\). The domain of definition of the function makes a difference now. A function that is not uniformly continuous on a larger set, may be uniformly continuous when restricted to a smaller set.

The function \(f \colon (0,1) \to {\mathbb{R}}\), defined by \(f(x) := \nicefrac{1}{x}\) is not uniformly continuous, but it is continuous.

Proof: Given \(\epsilon > 0\), then for \(\epsilon > \left\lvert {\nicefrac{1}{x}-\nicefrac{1}{y}} \right\rvert\) to hold we must have \[\epsilon > \left\lvert {\nicefrac{1}{x}-\nicefrac{1}{y}} \right\rvert = \frac{\left\lvert {y-x} \right\rvert}{\left\lvert {xy} \right\rvert} = \frac{\left\lvert {y-x} \right\rvert}{xy} ,\] or \[\left\lvert {x-y} \right\rvert < xy \epsilon .\] Therefore, to satisfy the definition of uniform continuity we would have to have \(\delta \leq xy \epsilon\) for all \(x,y\) in \((0,1)\), but that would mean that \(\delta \leq 0\). Therefore there is no single \(\delta > 0\).

\(f \colon [0,1] \to {\mathbb{R}}\), defined by \(f(x) := x^2\) is uniformly continuous.

Proof: Note that \(0 \leq x,c \leq 1\). Then \[\left\lvert {x^2-c^2} \right\rvert = \left\lvert {x+c} \right\rvert\left\lvert {x-c} \right\rvert \leq (\left\lvert {x} \right\rvert+\left\lvert {c} \right\rvert) \left\lvert {x-c} \right\rvert \leq (1+1)\left\lvert {x-c} \right\rvert .\] Therefore given \(\epsilon > 0\), let \(\delta := \nicefrac{\epsilon}{2}\). If \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {x^2-c^2} \right\rvert < \epsilon\).

On the other hand, \(f \colon {\mathbb{R}}\to {\mathbb{R}}\), defined by \(f(x) := x^2\) is not uniformly continuous.

Proof: Suppose it is uniformly continuous, then for all \(\epsilon > 0\), there would exist a \(\delta > 0\) such that if \(\left\lvert {x-c} \right\rvert < \delta\), then \(\left\lvert {x^2 -c^2} \right\rvert < \epsilon\). Take \(x > 0\) and let \(c := x+\nicefrac{\delta}{2}\). Write \[\epsilon > \left\lvert {x^2-c^2} \right\rvert = \left\lvert {x+c} \right\rvert\left\lvert {x-c} \right\rvert = (2x+\nicefrac{\delta}{2})\nicefrac{\delta}{2} \geq \delta x .\] Therefore \(x < \nicefrac{\epsilon}{\delta}\) for all \(x > 0\), which is a contradiction.

We have seen that if \(f\) is defined on an interval that is either not closed or not bounded, then \(f\) can be continuous, but not uniformly continuous. For a closed and bounded interval \([a,b]\), we can, however, make the following statement.

[unifcont:thm] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Then \(f\) is uniformly continuous.

We prove the statement by contrapositive. Suppose \(f\) is not uniformly continuous. We will prove that there is some \(c \in [a,b]\) where \(f\) is not continuous. Let us negate the definition of uniformly continuous. There exists an \(\epsilon > 0\) such that for every \(\delta > 0\), there exist points \(x, y\) in \(S\) with \(\left\lvert {x-y} \right\rvert < \delta\) and \(\left\lvert {f(x)-f(y)} \right\rvert \geq \epsilon\).

So for the \(\epsilon > 0\) above, we find sequences \(\{ x_n \}\) and \(\{ y_n \}\) such that \(\left\lvert {x_n-y_n} \right\rvert < \nicefrac{1}{n}\) and such that \(\left\lvert {f(x_n)-f(y_n)} \right\rvert \geq \epsilon\). By , there exists a convergent subsequence \(\{ x_{n_k} \}\). Let \(c := \lim\, x_{n_k}\). As \(a \leq x_{n_k} \leq b\), then \(a \leq c \leq b\). Write \[\left\lvert {y_{n_k} - c} \right\rvert = \left\lvert {y_{n_k} - x_{n_k} + x_{n_k} - c} \right\rvert \leq \left\lvert {y_{n_k} - x_{n_k}} \right\rvert + \left\lvert {x_{n_k}-c} \right\rvert < \nicefrac{1}{n_k} + \left\lvert {x_{n_k}-c} \right\rvert .\] As \(\nicefrac{1}{n_k}\) and \(\left\lvert {x_{n_k}-c} \right\rvert\) both go to zero when \(k\) goes to infinity, \(\{ y_{n_k} \}\) converges and the limit is \(c\). We now show that \(f\) is not continuous at \(c\). We estimate \[\begin{split} \left\lvert {f(x_{n_k}) - f(c)} \right\rvert & = \left\lvert {f(x_{n_k}) - f(y_{n_k}) + f(y_{n_k}) - f(c)} \right\rvert \\ & \geq \left\lvert {f(x_{n_k}) - f(y_{n_k})} \right\rvert - \left\lvert {f(y_{n_k}) - f(c)} \right\rvert \\ & \geq \epsilon - \left\lvert {f(y_{n_k})-f(c)} \right\rvert . \end{split}\] Or in other words \[\left\lvert {f(x_{n_k})-f(c)} \right\rvert + \left\lvert {f(y_{n_k})-f(c)} \right\rvert \geq \epsilon .\] At least one of the sequences \(\{ f(x_{n_k}) \}\) or \(\{ f(y_{n_k}) \}\) cannot converge to \(f(c)\), otherwise the left hand side of the inequality would go to zero while the right-hand side is positive. Thus \(f\) cannot be continuous at \(c\).

#### Continuous extension

Before we get to continuous extension, we show the following useful lemma. It says that uniformly continuous functions behave nicely with respect to Cauchy sequences. The new issue here is that for a Cauchy sequence we no longer know where the limit ends up; it may not end up in the domain of the function.

[unifcauchycauchy:lemma] Let \(f \colon S \to {\mathbb{R}}\) be a uniformly continuous function. Let \(\{ x_n \}\) be a Cauchy sequence in \(S\). Then \(\{ f(x_n) \}\) is Cauchy.

Let \(\epsilon > 0\) be given. Then there is a \(\delta > 0\) such that \(\left\lvert {f(x)-f(y)} \right\rvert < \epsilon\) whenever \(\left\lvert {x-y} \right\rvert < \delta\). Now find an \(M \in {\mathbb{N}}\) such that for all \(n, k \geq M\) we have \(\left\lvert {x_n-x_k} \right\rvert < \delta\). Then for all \(n, k \geq M\) we have \(\left\lvert {f(x_n)-f(x_k)} \right\rvert < \epsilon\).

An application of the above lemma is the following theorem. It says that a function on an open interval is uniformly continuous if and only if it can be extended to a continuous function on the closed interval.

[context:thm] A function \(f \colon (a,b) \to {\mathbb{R}}\) is uniformly continuous if and only if the limits \[L_a := \lim_{x \to a} f(x) \qquad \text{and} \qquad L_b := \lim_{x \to b} f(x)\] exist and the function \(\widetilde{f} \colon [a,b] \to {\mathbb{R}}\) defined by \[\widetilde{f}(x) := \begin{cases} f(x) & \text{ if $x \in (a,b)$,} \\ L_a & \text{ if $x = a$,} \\ L_b & \text{ if $x = b$,} \end{cases}\] is continuous.

One direction is not hard to prove. If \(\widetilde{f}\) is continuous, then it is uniformly continuous by . As \(f\) is the restriction of \(\widetilde{f}\) to \((a,b)\), then \(f\) is also uniformly continuous (easy exercise).

Now suppose \(f\) is uniformly continuous. We must first show that the limits \(L_a\) and \(L_b\) exist. Let us concentrate on \(L_a\). Take a sequence \(\{ x_n \}\) in \((a,b)\) such that \(\lim\, x_n = a\). The sequence is a Cauchy sequence and hence by , the sequence \(\{ f(x_n) \}\) is Cauchy and therefore convergent. We have some number \(L_1 := \lim\, f(x_n)\). Take another sequence \(\{ y_n \}\) in \((a,b)\) such that \(\lim\, y_n = a\). By the same reasoning we get \(L_2 := \lim\, f(y_n)\). If we show that \(L_1 = L_2\), then the limit \(L_a = \lim_{x\to a} f(x)\) exists. Let \(\epsilon > 0\) be given, find \(\delta > 0\) such that \(\left\lvert {x-y} \right\rvert < \delta\) implies \(\left\lvert {f(x)-f(y)} \right\rvert < \nicefrac{\epsilon}{3}\). Find \(M \in {\mathbb{N}}\) such that for \(n \geq M\) we have \(\left\lvert {a-x_n} \right\rvert < \nicefrac{\delta}{2}\), \(\left\lvert {a-y_n} \right\rvert < \nicefrac{\delta}{2}\), \(\left\lvert {f(x_n)-L_1} \right\rvert < \nicefrac{\epsilon}{3}\), and \(\left\lvert {f(y_n)-L_2} \right\rvert < \nicefrac{\epsilon}{3}\). Then for \(n \geq M\) we have \[\left\lvert {x_n-y_n} \right\rvert = \left\lvert {x_n-a+a-y_n} \right\rvert \leq \left\lvert {x_n-a} \right\rvert+\left\lvert {a-y_n} \right\rvert < \nicefrac{\delta}{2} + \nicefrac{\delta}{2} = \delta.\] So \[\begin{split} \left\lvert {L_1-L_2} \right\rvert &= \left\lvert {L_1-f(x_n)+f(x_n)-f(y_n)+f(y_n)-L_2} \right\rvert \\ & \leq \left\lvert {L_1-f(x_n)} \right\rvert+\left\lvert {f(x_n)-f(y_n)} \right\rvert+\left\lvert {f(y_n)-L_2} \right\rvert \\ & \leq \nicefrac{\epsilon}{3} + \nicefrac{\epsilon}{3} + \nicefrac{\epsilon}{3} = \epsilon . \end{split}\] Therefore \(L_1 = L_2\). Thus \(L_a\) exists. To show that \(L_b\) exists is left as an exercise.

Now that we know that the limits \(L_a\) and \(L_b\) exist, we are done. If \(\lim_{x\to a} f(x)\) exists, then \(\lim_{x\to a} \widetilde{f}(x)\) exists (See ). Similarly with \(L_b\). Hence \(\widetilde{f}\) is continuous at \(a\) and \(b\). And since \(f\) is continuous at \(c \in (a,b)\), then \(\widetilde{f}\) is continuous at \(c \in (a,b)\).

#### Lipschitz continuous functions

Let \(f \colon S \to {\mathbb{R}}\) be a function such that there exists a number \(K\) such that for all \(x\) and \(y\) in \(S\) we have \[\left\lvert {f(x)-f(y)} \right\rvert \leq K \left\lvert {x-y} \right\rvert .\] Then \(f\) is said to be *Lipschitz continuous* ^{16}.

A large class of functions is Lipschitz continuous. Be careful, just as for uniformly continuous functions, the domain of definition of the function is important. See the examples below and the exercises. First we justify the use of the word *continuous*.

A Lipschitz continuous function is uniformly continuous.

Let \(f \colon S \to {\mathbb{R}}\) be a function and let \(K\) be a constant such that for all \(x, y\) in \(S\) we have \(\left\lvert {f(x)-f(y)} \right\rvert \leq K \left\lvert {x-y} \right\rvert\).

Let \(\epsilon > 0\) be given. Take \(\delta := \nicefrac{\epsilon}{K}\). For any \(x\) and \(y\) in \(S\) such that \(\left\lvert {x-y} \right\rvert < \delta\) we have that \[\left\lvert {f(x)-f(y)} \right\rvert \leq K \left\lvert {x-y} \right\rvert < K \delta = K \frac{\epsilon}{K} = \epsilon .\] Therefore \(f\) is uniformly continuous.

We interpret Lipschitz continuity geometrically. If \(f\) is a Lipschitz continuous function with some constant \(K\). We rewrite the inequality to say that for \(x \not=y\) we have \[\left\lvert {\frac{f(x)-f(y)}{x-y}} \right\rvert \leq K .\] The quantity \(\frac{f(x)-f(y)}{x-y}\) is the slope of the line between the points \(\bigl(x,f(x)\bigr)\) and \(\bigl(y,f(y)\bigr)\), that is, a *secant line*. Therefore, \(f\) is Lipschitz continuous if and only if every line that intersects the graph of \(f\) in at least two distinct points has slope less than or equal to \(K\). See .

The functions \(\sin(x)\) and \(\cos(x)\) are Lipschitz continuous. We have seen () the following two inequalities. \[\left\lvert {\sin(x)-\sin(y)} \right\rvert \leq \left\lvert {x-y} \right\rvert \qquad \text{and} \qquad \left\lvert {\cos(x)-\cos(y)} \right\rvert \leq \left\lvert {x-y} \right\rvert .\]

Hence sin and cos are Lipschitz continuous with \(K=1\).

The function \(f \colon [1,\infty) \to {\mathbb{R}}\) defined by \(f(x) := \sqrt{x}\) is Lipschitz continuous. Proof: \[\left\lvert {\sqrt{x}-\sqrt{y}} \right\rvert = \left\lvert {\frac{x-y}{\sqrt{x}+\sqrt{y}}} \right\rvert = \frac{\left\lvert {x-y} \right\rvert}{\sqrt{x}+\sqrt{y}} .\] As \(x \geq 1\) and \(y \geq 1\), we see that \(\frac{1}{\sqrt{x}+\sqrt{y}} \leq \frac{1}{2}\). Therefore \[\left\lvert {\sqrt{x}-\sqrt{y}} \right\rvert = \left\lvert {\frac{x-y}{\sqrt{x}+\sqrt{y}}} \right\rvert \leq \frac{1}{2} \left\lvert {x-y} \right\rvert.\]

On the other hand \(f \colon [0,\infty) \to {\mathbb{R}}\) defined by \(f(x) := \sqrt{x}\) is not Lipschitz continuous. Let us see why: Suppose we have \[\left\lvert {\sqrt{x}-\sqrt{y}} \right\rvert \leq K \left\lvert {x-y} \right\rvert ,\] for some \(K\). Let \(y=0\) to obtain \(\sqrt{x} \leq K x\). If \(K > 0\), then for \(x > 0\) we then get \(\nicefrac{1}{K} \leq \sqrt{x}\). This cannot possibly be true for all \(x > 0\). Thus no such \(K > 0\) exists and \(f\) is not Lipschitz continuous.

The last example is a function that is uniformly continuous but not Lipschitz continuous. To see that \(\sqrt{x}\) is uniformly continuous on \([0,\infty)\) note that it is uniformly continuous on \([0,1]\) by . It is also Lipschitz (and therefore uniformly continuous) on \([1,\infty)\). It is not hard (exercise) to show that this means that \(\sqrt{x}\) is uniformly continuous on \([0,\infty)\).

## Exercises

Let \(f \colon S \to {\mathbb{R}}\) be uniformly continuous. Let \(A \subset S\). Then the restriction \(f|_A\) is uniformly continuous.

Let \(f \colon (a,b) \to {\mathbb{R}}\) be a uniformly continuous function. Finish the proof of by showing that the limit \(\lim\limits_{x \to b} f(x)\) exists.

Show that \(f \colon (c,\infty) \to {\mathbb{R}}\) for some \(c > 0\) and defined by \(f(x) := \nicefrac{1}{x}\) is Lipschitz continuous.

Show that \(f \colon (0,\infty) \to {\mathbb{R}}\) defined by \(f(x) := \nicefrac{1}{x}\) is not Lipschitz continuous.

Let \(A, B\) be intervals. Let \(f \colon A \to {\mathbb{R}}\) and \(g \colon B \to {\mathbb{R}}\) be uniformly continuous functions such that \(f(x) = g(x)\) for \(x \in A \cap B\). Define the function \(h \colon A \cup B \to {\mathbb{R}}\) by \(h(x) := f(x)\) if \(x \in A\) and \(h(x) := g(x)\) if \(x \in B \setminus A\). a) Prove that if \(A \cap B \not= \emptyset\), then \(h\) is uniformly continuous. b) Find an example where \(A \cap B = \emptyset\) and \(h\) is not even continuous.

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be a polynomial of degree \(d \geq 2\). Show that \(f\) is not Lipschitz continuous.

Let \(f \colon (0,1) \to {\mathbb{R}}\) be a bounded continuous function. Show that the function \(g(x) := x(1-x)f(x)\) is uniformly continuous.

Show that \(f \colon (0,\infty) \to {\mathbb{R}}\) defined by \(f(x) := \sin (\nicefrac{1}{x})\) is not uniformly continuous.

Let \(f \colon {\mathbb{Q}}\to {\mathbb{R}}\) be a uniformly continuous function. Show that there exists a uniformly continuous function \(\widetilde{f} \colon {\mathbb{R}}\to {\mathbb{R}}\) such that \(f(x) = \widetilde{f}(x)\) for all \(x \in {\mathbb{Q}}\).

a) Find a continuous \(f \colon (0,1) \to {\mathbb{R}}\) and a sequence \(\{ x_n \}\) in \((0,1)\) that is Cauchy, but such that \(\{ f(x_n) \}\) is not Cauchy. b) Prove that if \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is continuous, and \(\{ x_n \}\) is Cauchy, then \(\{ f(x_n) \}\) is Cauchy.

a) If \(f \colon S \to {\mathbb{R}}\) and \(g \colon S \to {\mathbb{R}}\) are uniformly continuous, then show that \(h \colon S \to {\mathbb{R}}\) given by \(h(x) := f(x) + g(x)\) is uniformly continuous.

b) If \(f \colon S \to {\mathbb{R}}\) is uniformly continuous and \(a \in {\mathbb{R}}\), then show that \(h \colon S \to {\mathbb{R}}\) given by \(h(x) := a f(x)\) is uniformly continuous.

a) If \(f \colon S \to {\mathbb{R}}\) and \(g \colon S \to {\mathbb{R}}\) are Lipschitz, then show that \(h \colon S \to {\mathbb{R}}\) given by \(h(x) := f(x) + g(x)\) is Lipschitz.

b) If \(f \colon S \to {\mathbb{R}}\) is Lipschitz and \(a \in {\mathbb{R}}\), then show that \(h \colon S \to {\mathbb{R}}\) given by \(h(x) := a f(x)\) is Lipschitz.

a) If \(f \colon [0,1] \to {\mathbb{R}}\) is given by \(f(x) := x^m\) for an integer \(m \geq 0\), show \(f\) is Lipschitz and find the best (the smallest) Lipschitz constant \(K\) (depending on \(m\) of course). Hint: \((x-y)(x^{m-1} + x^{m-2}y + x^{m-3}y^2 + \cdots + x y^{m-2} + y^{m-1}) = x^m - y^m\).

b) Using the previous exercise, show that if \(f \colon [0,1] \to {\mathbb{R}}\) is a polynomial, that is, \(f(x) := a_m x^m + a_{m-1} x^{m-1} + \cdots + a_0\), then \(f\) is Lipschitz.

Suppose for \(f \colon [0,1] \to {\mathbb{R}}\) we have \(\left\lvert {f(x)-f(y)} \right\rvert \leq K \left\lvert {x-y} \right\rvert\), and \(f(0) = f(1) = 0\). Prove that \(\left\lvert {f(x)} \right\rvert \leq \nicefrac{K}{2}\). Further show by example that \(\nicefrac{K}{2}\) is the best possible, that is, there exists such a continuous function for which \(\left\lvert {f(x)} \right\rvert = \nicefrac{K}{2}\) for some \(x \in [0,1]\).

### Limits at infinity

*Note: less than 1 lecture (optional, can safely be omitted unless or is also covered)*

#### Limits at infinity

As for sequences, a continuous variable can also approach infinity. Let us make this notion precise.

We say \(\infty\) is a cluster point of \(S \subset {\mathbb{R}}\), if for every \(M \in {\mathbb{R}}\), there exists an \(x \in S\) such that \(x \geq M\). Similarly \(- \infty\) is a cluster point of \(S \subset {\mathbb{R}}\), if for every \(M \in {\mathbb{R}}\), there exists an \(x \in S\) such that \(x \leq M\).

Let \(f \colon S \to {\mathbb{R}}\) be a function, where \(\infty\) is a cluster point of \(S\). If there exists an \(L \in {\mathbb{R}}\) such that for every \(\epsilon > 0\), there is an \(M \in {\mathbb{R}}\) such that \[\left\lvert {f(x) - L} \right\rvert < \epsilon\] whenever \(x \geq M\), then we say \(f(x)\) *converges* to \(L\) as \(x\) goes to \(\infty\). We call \(L\) the *limit* and write \[\lim_{x \to \infty} f(x) := L .\] Alternatively we write \(f(x) \to L\) as \(x \to \infty\).

Similarly, if \(-\infty\) is a cluster point of \(S\) and there exists an \(L \in {\mathbb{R}}\) such that for every \(\epsilon > 0\), there is an \(M \in {\mathbb{R}}\) such that \[\left\lvert {f(x) - L} \right\rvert < \epsilon\] whenever \(x \leq M\), then we say \(f(x)\) *converges* to \(L\) as \(x\) goes to \(-\infty\). We call \(L\) the *limit* and write \[\lim_{x \to -\infty} f(x) := L .\] Alternatively we write \(f(x) \to L\) as \(x \to -\infty\).

We cheated a little bit again and said *the* limit. We leave it as an exercise for the reader to prove the following proposition.

[liminfty:unique] The limit at \(\infty\) or \(-\infty\) as defined above is unique if it exists.

Let \(f(x) := \frac{1}{\left\lvert {x} \right\rvert+1}\). Then \[\lim_{x\to \infty} f(x) = 0 \qquad \text{and} \qquad \lim_{x\to -\infty} f(x) = 0 .\]

Proof: Let \(\epsilon > 0\) be given. Find \(M > 0\) large enough so that \(\frac{1}{M+1} < \epsilon\). If \(x \geq M\), then \(\frac{1}{x+1} \leq \frac{1}{M+1} < \epsilon\). Since \(\frac{1}{\left\lvert {x} \right\rvert+1} > 0\) for all \(x\) the first limit is proved. The proof for \(-\infty\) is left to the reader.

Let \(f(x) := \sin(\pi x)\). Then \(\lim_{x\to\infty} f(x)\) does not exist. To prove this fact note that if \(x = 2n+\nicefrac{1}{2}\) for some \(n \in {\mathbb{N}}\) then \(f(x)=1\), while if \(x = 2n+\nicefrac{3}{2}\) then \(f(x)=-1\), so they cannot both be within a small \(\epsilon\) of a single real number.

We must be careful not to confuse continuous limits with limits of sequences. For \(f(x) = \sin(\pi x)\) we could say \[\lim_{n \to \infty} f(n) = 0, \qquad \text{but} \qquad \lim_{x \to \infty} f(x) ~ \text{does not exist}.\] Of course the notation is ambiguous. We are simply using the convention that \(n \in {\mathbb{N}}\), while \(x \in {\mathbb{R}}\). When the notation is not clear, it is good to explicitly mention where the variable lives, or what kind of limit are you using.

There is a connection of continuous limits to limits of sequences, but we must take all sequences going to infinity, just as before in .

[seqflimitinf:lemma] Suppose \(f \colon S \to {\mathbb{R}}\) is a function, \(\infty\) is a cluster point of \(S \subset {\mathbb{R}}\), and \(L \in {\mathbb{R}}\). Then \[\lim_{x\to\infty} f(x) = L % \qquad \text{if and only if} \qquad\] if and only if \[\lim_{n\to\infty} f(x_n) = L% ~~\text{for all sequences $\{ x_n \}$ such that $\lim\, x_n = \infty$} .\] for all sequences \(\{ x_n \}\) such that \(\lim\limits_{n\to\infty} x_n = \infty\).

The lemma holds for the limit as \(x \to -\infty\). Its proof is almost identical and is left as an exercise.

First suppose \(f(x) \to L\) as \(x \to \infty\). Given an \(\epsilon > 0\), there exists an \(M\) such that for all \(x \geq M\) we have \(\left\lvert {f(x)-L} \right\rvert < \epsilon\). Let \(\{ x_n \}\) be a sequence in \(S\) such that \(\lim \, x_n = \infty\). Then there exists an \(N\) such that for all \(n \geq N\) we have \(x_n \geq M\). And thus \(\left\lvert {f(x_n)-L} \right\rvert < \epsilon\).

We prove the converse by contrapositive. Suppose \(f(x)\) does not go to \(L\) as \(x \to \infty\). This means that there exists an \(\epsilon > 0\), such that for every \(M \in {\mathbb{N}}\), there exists an \(x \in S\), \(x \geq M\), let us call it \(x_M\), such that \(\left\lvert {f(x_M)-L} \right\rvert \geq \epsilon\). Consider the sequence \(\{ x_n \}\). Clearly \(\{ f(x_n) \}\) does not converge to \(L\). It remains to note that \(\lim\, x_n = \infty\), because \(x_n \geq n\) for all \(n\).

Using the lemma, we again translate results about sequential limits into results about continuous limits as \(x\) goes to infinity. That is, we have almost immediate analogues of the corollaries in . We simply allow the cluster point \(c\) to be either \(\infty\) or \(-\infty\), in addition to a real number. We leave it to the student to verify these statements.

#### Infinite limit

Just as for sequences, it is often convenient to distinguish certain divergent sequences, and talk about limits being infinite almost as if the limits existed.

Let \(f \colon S \to {\mathbb{R}}\) be a function and suppose \(S\) has \(\infty\) as a cluster point. We say \(f(x)\) *diverges to infinity* as \(x\) goes to \(\infty\), if for every \(N \in {\mathbb{R}}\) there exists an \(M \in {\mathbb{R}}\) such that \[f(x) > N\] whenever \(x \in S\) and \(x \geq M\). We write \[\lim_{x \to \infty} f(x) := \infty ,\] or we say that \(f(x) \to \infty\) as \(x \to \infty\).

A similar definition can be made for limits as \(x \to -\infty\) or as \(x \to c\) for a finite \(c\). Also similar definitions can be made for limits being \(-\infty\). Stating these definitions is left as an exercise. Note that sometimes *converges to infinity* is used. We can again use sequential limits, and an analogue of is left as an exercise.

Let us show that \(\lim_{x \to \infty} \frac{1+x^2}{1+x} = \infty\).

Proof: For \(x \geq 1\) we have \[\frac{1+x^2}{1+x} \geq \frac{x^2}{x+x} = \frac{x}{2} .\] Given \(N \in {\mathbb{R}}\), take \(M = \max \{ 2N+1 , 1 \}\). If \(x \geq M\), then \(x \geq 1\) and \(\nicefrac{x}{2} > N\). So \[\frac{1+x^2}{1+x} \geq \frac{x}{2} > N .\]

#### Compositions

Finally, just as for limits at finite numbers we can compose functions easily.

[prop:inflimcompositions] Suppose \(f \colon A \to B\), \(g \colon B \to {\mathbb{R}}\), \(A, B \subset {\mathbb{R}}\), \(a \in {\mathbb{R}}\cup \{ -\infty, \infty\}\) is a cluster point of \(A\), and \(b \in {\mathbb{R}}\cup \{ -\infty, \infty\}\) is a cluster point of \(B\). Suppose \[\lim_{x \to a} f(x) = b\qquad \text{and} \qquad \lim_{y \to b} g(y) = c\] for some \(c \in {\mathbb{R}}\cup \{ -\infty, \infty \}\). If \(b \in B\), then suppose \(g(b) = c\). Then \[\lim_{x \to a} g\bigl(f(x)\bigr) = c .\]

The proof is straightforward, and left as an exercise. We already know the proposition when \(a, b, c \in {\mathbb{R}}\), see Exercises [exercise:contlimitcomposition] and [exercise:contlimitbadcomposition]. Again the requirement that \(g\) is continuous at \(b\), if \(b \in B\), is necessary.

Let \(h(x) := e^{-x^2+x}\). Then \[\lim_{x\to \infty} h(x) = 0 .\]

Proof: The claim follows once we know \[\lim_{x\to \infty} -x^2+x = -\infty\] and \[\lim_{y\to -\infty} e^y = 0 ,\] which is usually proved when the exponential function is defined.

## Exercises

Prove .

Let \(f \colon [1,\infty) \to {\mathbb{R}}\) be a function. Define \(g \colon (0,1] \to {\mathbb{R}}\) via \(g(x) := f(\nicefrac{1}{x})\). Using the definitions of limits directly, show that \(\lim_{x\to 0^+} g(x)\) exists if and only if \(\lim_{x\to \infty} f(x)\) exists, in which case they are equal.

Prove .

Let us justify terminology. Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be a function such that \(\lim_{x \to \infty} f(x) = \infty\) (diverges to infinity). Show that \(f(x)\) diverges (i.e. does not converge) as \(x \to \infty\).

Come up with the definitions for limits of \(f(x)\) going to \(-\infty\) as \(x \to \infty\), \(x \to -\infty\), and as \(x \to c\) for a finite \(c \in {\mathbb{R}}\). Then state the definitions for limits of \(f(x)\) going to \(\infty\) as \(x \to -\infty\), and as \(x \to c\) for a finite \(c \in {\mathbb{R}}\).

Suppose \(P(x) := x^n + a_{n-1} x^{n-1} + \cdots + a_1 x + a_0\) is a *monic polynomial* of degree \(n \geq 1\) (monic means that the coefficient of \(x^n\) is 1). a) Show that if \(n\) is even then \(\lim_{x\to\infty} P(x) = \lim_{x\to-\infty} P(x) = \infty\). b) Show that if \(n\) is odd then \(\lim_{x\to\infty} P(x) = \infty\) and \(\lim_{x\to-\infty} P(x) = -\infty\) (see previous exercise).

Let \(\{ x_n \}\) be a sequence. Consider \(S := {\mathbb{N}}\subset {\mathbb{R}}\), and \(f \colon S \to {\mathbb{R}}\) defined by \(f(n) := x_n\). Show that the two notions of limit, \[\lim_{n\to\infty} x_n \qquad \text{and} \qquad \lim_{x\to\infty} f(x)\] are equivalent. That is, show that if one exists so does the other one, and in this case they are equal.

Extend as follows. Suppose \(S \subset {\mathbb{R}}\) has a cluster point \(c \in {\mathbb{R}}\), \(c = \infty\), or \(c = -\infty\). Let \(f \colon S \to {\mathbb{R}}\) be a function and let \(L = \infty\) or \(L = -\infty\). Show that \[\lim_{x\to c} f(x) = L \qquad \text{if and only if} \qquad \lim_{n\to\infty} f(x_n) = L ~~\text{for all sequences $\{ x_n \}$ such that $\lim\, x_n = c$} .\]

### Monotone functions and continuity

*Note: 1 lecture (optional, can safely be omitted unless is also covered, requires )*

Let \(S \subset {\mathbb{R}}\). We say \(f \colon S \to {\mathbb{R}}\) is *increasing* (resp. *strictly increasing*) if \(x,y \in S\) with \(x < y\) implies \(f(x) \leq f(y)\) (resp. \(f(x) < f(y)\)). We define *decreasing* and *strictly decreasing* in the same way by switching the inequalities for \(f\).

If a function is either increasing or decreasing we say it is *monotone*. If it is strictly increasing or strictly decreasing we say it is *strictly monotone*.

Sometimes *nondecreasing* (resp. *nonincreasing*) is used for increasing (resp. decreasing) function to emphasize it is not strictly increasing (resp. strictly decreasing).

#### Continuity of monotone functions

It is easy to compute one-sided limits for monotone functions.

[prop:monotlimits] Let \(S \subset {\mathbb{R}}\), \(c \in {\mathbb{R}}\), and \(f \colon S \to {\mathbb{R}}\) be increasing. If \(c\) is a cluster point of \(S \cap (-\infty,c)\), then \[\lim_{x \to c^-} f(x) = \sup \{ f(x) : x < c, x \in S \} ,\] and if \(c\) is a cluster point of \(S \cap (c,\infty)\), then \[\lim_{x \to c^+} f(x) = \inf \{ f(x) : x > c, x \in S \} .\]

Similarly, if \(f\) is decreasing and \(c\) is a cluster point of \(S \cap (-\infty,c)\), then \[\lim_{x \to c^-} f(x) = \inf \{ f(x) : x < c, x \in S \} ,\] and if \(c\) is a cluster point of \(S \cap (c,\infty)\), then \[\lim_{x \to c^+} f(x) = \sup \{ f(x) : x > c, x \in S \} .\]

In particular all the one-sided limits exist whenever they make sense. If we from now on say that say a the left hand limit \(x \to c^-\) exists we mean that \(c\) is a cluster point of \(S \cap (-\infty,c)\).

Let us assume \(f\) is increasing, and we will show the first equality. The rest of the proof is very similar and is left as an exercise.

Let \(a := \sup \{ f(x) : x < c, x \in S \}\). If \(a = \infty\), then for every \(M \in {\mathbb{R}}\), there exists an \(x_M\) such that \(f(x_M) > M\). As \(f\) is increasing, \(f(x) \geq f(x_M) > M\) for all \(x \in S\) with \(x > x_M\). If we take \(\delta = c-x_M\) we obtain the definition of the limit going to infinity.

So assume \(a < \infty\). Let \(\epsilon > 0\) be given. Because \(a\) is the supremum, there exists an \(x_\epsilon < c\), \(x_\epsilon \in S\), such that \(f(x_\epsilon) > a-\epsilon\). As \(f\) is increasing, if \(x \in S\) and \(x_\epsilon < x < c\), we have \(a-\epsilon < f(x_\epsilon) \leq f(x) \leq a\). Let \(\delta := c-x_\epsilon\). Then for \(x \in S \cap (-\infty,c)\) with \(\left\lvert {x-c} \right\rvert < \delta\), we have \(\left\lvert {f(x)-a} \right\rvert < \epsilon\).

Suppose \(f \colon S \to {\mathbb{R}}\), \(c \in S\) and that both one-sided limits exist. Since \(f(x) \leq f(c) \leq f(y)\) whenever \(x < c < y\), taking the limits we obtain \[\lim_{x \to c^-} f(x) \leq f(c) \leq \lim_{x \to c^+} f(x) .\] Then \(f\) is continuous at \(c\) if and only if both limits are equal to each other (and hence equal to \(f(c)\)). See also . See to get an idea of a what a discontinuity looks like.

[cor:continterval] If \(I \subset {\mathbb{R}}\) is an interval and \(f \colon I \to {\mathbb{R}}\) is monotone and not constant, then \(f(I)\) is an interval if and only if \(f\) is continuous.

Assuming \(f\) is not constant is to avoid the technicality that \(f(I)\) is a single point in that case; \(f(I)\) is a single point if and only if \(f\) is constant. A constant function is continuous.

If \(f\) is continuous then \(f(I)\) being an interval is a consequence of . See also .

Let us prove the reverse direction by contrapositive. Suppose \(f\) is not continuous at \(c \in I\), and that \(c\) is not an endpoint of \(I\). Without loss of generality suppose \(f\) is increasing. Let \[a := \lim_{x \to c^-} f(x) = \sup \{ f(x) : x \in I, x < c \} , \qquad b := \lim_{x \to c^+} f(x) = \inf \{ f(x) : x \in I, x > c \} .\] As \(c\) is a discontinuity, \(a < b\). If \(x < c\), then \(f(x) \leq a\), and if \(x > c\), then \(f(x) \geq b\). Therefore any point in \((a,b) \setminus \{ f(c) \}\) is not in \(f(I)\). However there exists \(x_1 \in S\), \(x_1 < c\) so \(f(x_1) \leq a\), and there exists \(x_2 \in S\), \(x_2 > c\) so \(f(x_2) \geq b\). Both \(f(x_1)\) and \(f(x_2)\) are in \(f(I)\), and so \(f(I)\) is not an interval. See .

When \(c \in I\) is an endpoint, the proof is similar and is left as an exercise.

A striking property of monotone functions is that they cannot have too many discontinuities.

[cor:monotcountcont] Let \(I \subset {\mathbb{R}}\) be an interval and \(f \colon I \to {\mathbb{R}}\) be monotone. Then \(f\) has at most countably many discontinuities.

Let \(E \subset I\) be the set of all discontinuities that are not endpoints of \(I\). As there are only two endpoints, it is enough to show that \(E\) is countable. Without loss of generality, suppose \(f\) is increasing. We will define an injection \(h \colon E \to {\mathbb{Q}}\). For each \(c \in E\) the one-sided limits of \(f\) both exist as \(c\) is not an endpoint. Let \[a := \lim_{x \to c^-} f(x) = \sup \{ f(x) : x \in I, x < c \} , \qquad b := \lim_{x \to c^+} f(x) = \inf \{ f(x) : x \in I, x > c \} .\] As \(c\) is a discontinuity, we have \(a < b\). There exists a rational number \(q \in (a,b)\), so let \(h(c) := q\). Because \(f\) is increasing, \(q\) cannot correspond to any other discontinuity, so after making this choice for all \(c \in E\), we have that \(h\) is one-to-one (injective). Therefore, \(E\) is countable.

[example:countdiscont] By \(\lfloor x \rfloor\) denote the largest integer less than or equal to \(x\). Define \(f \colon [0,1] \to {\mathbb{R}}\) by \[f(x) := x + \sum_{n=0}^{\lfloor 1/(1-x) \rfloor} 2^{-n} ,\] for \(x < 1\) and \(f(1) = 3\). It is left as an exercise to show that \(f\) is strictly increasing, bounded, and has a discontinuity at all points \(1-\nicefrac{1}{k}\) for \(k \in {\mathbb{N}}\). In particular, there are countably many discontinuities, but the function is bounded and defined on a closed bounded interval.

#### Continuity of inverse functions

A strictly monotone function \(f\) is one-to-one (injective). To see this notice that if \(x \not= y\) then we can assume \(x < y\). Then either \(f(x) < f(y)\) if \(f\) is strictly increasing or \(f(x) > f(y)\) if \(f\) is strictly decreasing, so \(f(x) \not= f(y)\). Hence, it must have an inverse \(f^{-1}\) defined on its range.

[prop:invcont] If \(I \subset {\mathbb{R}}\) is an interval and \(f \colon I \to {\mathbb{R}}\) is strictly monotone. Then the inverse \(f^{-1} \colon f(I) \to I\) is continuous.

Let us suppose \(f\) is strictly increasing. The proof is almost identical for a strictly decreasing function. Since \(f\) is strictly increasing, so is \(f^{-1}\). That is, if \(f(x) < f(y)\), then we must have \(x < y\) and therefore \(f^{-1}\bigl(f(x)\bigr) < f^{-1}\bigl(f(y)\bigr)\).

Take \(c \in f(I)\). If \(c\) is not a cluster point of \(f(I)\), then \(f^{-1}\) is continuous at \(c\) automatically. So let \(c\) be a cluster point of \(f(I)\). Suppose both of the following one-sided limits exist: \[\begin{aligned} x_0 & := \lim_{y \to c^-} f^{-1}(y) = \sup \{ f^{-1}(y) : y < c, y \in f(I) \} = \sup \{ x : f(x) < c, x \in I \} , \\ x_1 & := \lim_{y \to c^+} f^{-1}(y) = \inf \{ f^{-1}(y) : y > c, y \in f(I) \} = \inf \{ x : f(x) > c, x \in I \} .\end{aligned}\] We have \(x_0 \leq x_1\) as \(f^{-1}\) is increasing. For all \(x > x_0\) with \(x \in I\), we have \(f(x) \geq c\). As \(f\) is strictly increasing, we must have \(f(x) > c\) for all \(x > x_0\), \(x \in I\). Therefore, \[\{ x : x > x_0, x \in I \} \subset \{ x : f(x) > c, x \in I \}.\] The infimum of the left hand set is \(x_0\) and the infimum of the right hand set is \(x_1\), so we obtain \(x_0 \geq x_1\). So \(x_1 = x_0\), and \(f^{-1}\) is continuous at \(c\).

If one of the one-sided limits does not exist the argument is similar and is left as an exercise.

The proposition does not require \(f\) itself to be continuous. For example, let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) \[f(x) := \begin{cases} x & \text{if $x < 0$}, \\ x+1 & \text{if $x \geq 0$}. \\ \end{cases}\] The function \(f\) is not continuous at \(0\). The image of \(I = {\mathbb{R}}\) is the set \((-\infty,0)\cup [1,\infty)\), not an interval. Then \(f^{-1} \colon (-\infty,0)\cup [1,\infty) \to {\mathbb{R}}\) can be written as \[f^{-1}(x) = \begin{cases} x & \text{if $x < 0$}, \\ x-1 & \text{if $x \geq 1$}. \end{cases}\] It is not difficult to see that \(f^{-1}\) is a continuous function.

Notice what happens with the proposition if \(f(I)\) is an interval. In that case we could simply apply to both \(f\) and \(f^{-1}\). That is, if \(f \colon I \to J\) is an onto strictly monotone function and \(I\) and \(J\) are intervals, then both \(f\) and \(f^{-1}\) are continuous. Furthermore \(f(I)\) is an interval precisely when \(f\) is continuous.

## Exercises

Suppose \(f \colon [0,1] \to {\mathbb{R}}\) is monotone. Prove \(f\) is bounded.

Finish the proof of .

Finish the proof of .

Prove the claims in .

Finish the proof of .

Suppose \(S \subset {\mathbb{R}}\), and \(f \colon S \to {\mathbb{R}}\) is an increasing function. a) If \(c\) is a cluster point of \(S \cap (c,\infty)\) show that \(\lim\limits_{x\to c^+} f(x) < \infty\). b) If \(c\) is a cluster point of \(S \cap (-\infty,c)\) and \(\lim\limits_{x\to c^-} f(x) = \infty\), prove that \(S \subset (-\infty,c)\).

Suppose \(I \subset {\mathbb{R}}\) is an interval and \(f \colon I \to {\mathbb{R}}\) is a function such that for each \(c \in I\), there exist \(a, b \in {\mathbb{R}}\) with \(a > 0\) such that \(f(x) \geq a x + b\) for all \(x \in I\) and \(f(c) = a c + b\). Show that \(f\) is strictly increasing.

Suppose \(f \colon I \to J\) is a continuous, bijective (one-to-one and onto) function for two intervals \(I\) and \(J\). Show that \(f\) is strictly monotone.

Consider a monotone function \(f \colon I \to {\mathbb{R}}\) on an interval \(I\). Prove that there exists a function \(g \colon I \to {\mathbb{R}}\) such that \(\lim\limits_{x \to c^-} g(x) = g(c)\) for all \(c \in I\), except the smaller (left) endpoint of \(I\), and such that \(g(x) = f(x)\) for all but countably many \(x\).

a) Let \(S \subset {\mathbb{R}}\) be any subset. If \(f \colon S \to {\mathbb{R}}\) is increasing, then show that there exists an increasing \(F \colon {\mathbb{R}}\to {\mathbb{R}}\) such that \(f(x) = F(x)\) for all \(x \in S\). b) Find an example of a strictly increasing \(f \colon S \to {\mathbb{R}}\) such that an increasing \(F\) as above is never strictly increasing.

[exercise:increasingfuncdiscatQ] Find an example of an increasing function \(f \colon [0,1] \to {\mathbb{R}}\) that has a discontinuity at each rational number. Then show that the image \(f([0,1])\) contains no interval. Hint: Enumerate the rational numbers and define the function with a series.

## The Derivative

### The derivative

*Note: 1 lecture*

The idea of a derivative is the following. Let us suppose a graph of a function looks locally like a straight line. We can then talk about the slope of this line. The slope tells us the rate at which the value of the function changing at the particular point. Of course, we are leaving out any function that has corners or discontinuities. Let us be precise.

#### Definition and basic properties

Let \(I\) be an interval, let \(f \colon I \to {\mathbb{R}}\) be a function, and let \(c \in I\). If the limit \[L := \lim_{x \to c} \frac{f(x)-f(c)}{x-c}\] exists, then we say \(f\) is *differentiable* at \(c\), that \(L\) is the *derivative* of \(f\) at \(c\), and write \(f'(c) := L\).

If \(f\) is differentiable at all \(c \in I\), then we simply say that \(f\) is *differentiable*, and then we obtain a function \(f' \colon I \to {\mathbb{R}}\).

The expression \(\frac{f(x)-f(c)}{x-c}\) is called the *difference quotient*.

The graphical interpretation of the derivative is depicted in . The left-hand plot gives the line through \(\bigl(c,f(c)\bigr)\) and \(\bigl(x,f(x)\bigr)\) with slope \(\frac{f(x)-f(c)}{x-c}\), that is, the so-called *secant line*. When we take the limit as \(x\) goes to \(c\), we get the right-hand plot, where we see that the derivative of the function at the point \(c\) is the slope of the line tangent to the graph of \(f\) at the point \(\bigl(c,f(c)\bigr)\).

We allow \(I\) to be a closed interval and we allow \(c\) to be an endpoint of \(I\). Some calculus books do not allow \(c\) to be an endpoint of an interval, but all the theory still works by allowing it, and it makes our work easier.

Let \(f(x) := x^2\) defined on the whole real line. We find that \[\lim_{x\to c} \frac{x^2-c^2}{x-c} = \lim_{x\to c} \frac{(x+c)(x-c)}{x-c} = \lim_{x\to c} (x+c) = 2c.\] Therefore \(f'(c) = 2c\).

The function \(f(x) := \left\lvert {x} \right\rvert\) is not differentiable at the origin. When \(x > 0\), then \[\frac{\left\lvert {x} \right\rvert-\left\lvert {0} \right\rvert}{x-0} = \frac{x-0}{x-0} = 1 ,\] and when \(x < 0\) we have \[\frac{\left\lvert {x} \right\rvert-\left\lvert {0} \right\rvert}{x-0} = \frac{-x-0}{x-0} = -1 .\]

A famous example of Weierstrass shows that there exists a continuous function that is not differentiable at *any* point. The construction of this function is beyond the scope of this book. On the other hand, a differentiable function is always continuous.

Let \(f \colon I \to {\mathbb{R}}\) be differentiable at \(c \in I\), then it is continuous at \(c\).

We know the limits \[\lim_{x\to c}\frac{f(x)-f(c)}{x-c} = f'(c) \qquad \text{and} \qquad \lim_{x\to c}(x-c) = 0\] exist. Furthermore, \[f(x)-f(c) = \left( \frac{f(x)-f(c)}{x-c} \right) (x-c) .\] Therefore the limit of \(f(x)-f(c)\) exists and \[\lim_{x\to c} \bigl( f(x)-f(c) \bigr) = \left(\lim_{x\to c} \frac{f(x)-f(c)}{x-c} \right) \left(\lim_{x\to c} (x-c) \right) = f'(c) \cdot 0 = 0.\] Hence \(\lim\limits_{x\to c} f(x) = f(c)\), and \(f\) is continuous at \(c\).

An important property of the derivative is linearity. The derivative is the approximation of a function by a straight line. The slope of a line through two points changes linearly when the \(y\)-coordinates are changed linearly. By taking the limit, it makes sense that the derivative is linear.

Let \(I\) be an interval, let \(f \colon I \to {\mathbb{R}}\) and \(g \colon I \to {\mathbb{R}}\) be differentiable at \(c \in I\), and let \(\alpha \in {\mathbb{R}}\).

- Define \(h \colon I \to {\mathbb{R}}\) by \(h(x) := \alpha f(x)\). Then \(h\) is differentiable at \(c\) and \(h'(c) = \alpha f'(c)\).
- Define \(h \colon I \to {\mathbb{R}}\) by \(h(x) := f(x) + g(x)\). Then \(h\) is differentiable at \(c\) and \(h'(c) = f'(c) + g'(c)\).

First, let \(h(x) := \alpha f(x)\). For \(x \in I\), \(x \not= c\) we have \[\frac{h(x)-h(c)}{x-c} = \frac{\alpha f(x) - \alpha f(c)}{x-c} = \alpha \frac{f(x) - f(c)}{x-c} .\] The limit as \(x\) goes to \(c\) exists on the right by . We get \[\lim_{x\to c}\frac{h(x)-h(c)}{x-c} = \alpha \lim_{x\to c} \frac{f(x) - f(c)}{x-c} .\] Therefore \(h\) is differentiable at \(c\), and the derivative is computed as given.

Next, define \(h(x) := f(x)+g(x)\). For \(x \in I\), \(x \not= c\) we have \[\frac{h(x)-h(c)}{x-c} = \frac{\bigl(f(x) + g(x)\bigr) - \bigl(f(c) + g(c)\bigr)}{x-c} = \frac{f(x) - f(c)}{x-c} + \frac{g(x) - g(c)}{x-c} .\] The limit as \(x\) goes to \(c\) exists on the right by . We get \[\lim_{x\to c}\frac{h(x)-h(c)}{x-c} = \lim_{x\to c} \frac{f(x) - f(c)}{x-c} + \lim_{x\to c}\frac{g(x) - g(c)}{x-c} .\] Therefore \(h\) is differentiable at \(c\) and the derivative is computed as given.

It is not true that the derivative of a multiple of two functions is the multiple of the derivatives. Instead we get the so-called *product rule* or the *Leibniz rule* ^{17}.

Let \(I\) be an interval, let \(f \colon I \to {\mathbb{R}}\) and \(g \colon I \to {\mathbb{R}}\) be functions differentiable at \(c\). If \(h \colon I \to {\mathbb{R}}\) is defined by \[h(x) := f(x) g(x) ,\] then \(h\) is differentiable at \(c\) and \[h'(c) = f(c) g'(c) + f'(c) g(c) .\]

The proof of the product rule is left as an exercise. The key is to use the identity \(f(x) g(x) - f(c) g(c) = f(x)\bigl( g(x) - g(c) \bigr) + g(c) \bigl( f(x) - f(c) \bigr)\).

Let \(I\) be an interval, let \(f \colon I \to {\mathbb{R}}\) and \(g \colon I \to {\mathbb{R}}\) be differentiable at \(c\) and \(g(x) \not= 0\) for all \(x \in I\). If \(h \colon I \to {\mathbb{R}}\) is defined by \[h(x) := \frac{f(x)}{g(x)},\] then \(h\) is differentiable at \(c\) and \[h'(c) = \frac{f'(c) g(c) - f(c) g'(c)}{{\bigl(g(c)\bigr)}^2} .\]

Again the proof is left as an exercise.

#### Chain rule

A useful rule for computing derivatives is the chain rule.

Let \(I_1, I_2\) be intervals, let \(g \colon I_1 \to I_2\) be differentiable at \(c \in I_1\), and \(f \colon I_2 \to {\mathbb{R}}\) be differentiable at \(g(c)\). If \(h \colon I_1 \to {\mathbb{R}}\) is defined by \[h(x) := (f \circ g) (x) = f\bigl(g(x)\bigr) ,\] then \(h\) is differentiable at \(c\) and \[h'(c) = f'\bigl(g(c)\bigr)g'(c) .\]

Let \(d := g(c)\). Define \(u \colon I_2 \to {\mathbb{R}}\) and \(v \colon I_1 \to {\mathbb{R}}\) by \[\begin{aligned} & u(y) := \begin{cases} \frac{f(y) - f(d)}{y-d} & \text{ if $y \not=d$,} \\ f'(d) & \text{ if $y = d$,} \end{cases} \\ & v(x) := \begin{cases} \frac{g(x) - g(c)}{x-c} & \text{ if $x \not=c$,} \\ g'(c) & \text{ if $x = c$.} \end{cases}\end{aligned}\] We note that \[f(y)-f(d) = u(y) (y-d) \qquad \text{and} \qquad g(x)-g(c) = v(x) (x-c) .\] We plug in to obtain \[h(x)-h(c) = f\bigl(g(x)\bigr)-f\bigl(g(c)\bigr) = u\bigl( g(x) \bigr) \bigl(g(x)-g(c)\bigr) = u\bigl( g(x) \bigr) \bigl(v(x) (x-c)\bigr) .\] Therefore, \[\label{eq:chainruleeq} \frac{h(x)-h(c)}{x-c} = u\bigl( g(x) \bigr) v(x) .\] We compute the limits \(\lim_{y \to d} u(y) = f'(d) = f'\bigl(g(c)\bigr)\) and \(\lim_{x \to c} v(x) = g'(c)\). That is, the functions \(u\) and \(v\) are continuous at \(d = g(c)\) and \(c\) respectively. Furthermore the function \(g\) is continuous at \(c\). Hence the limit of the right-hand side of [eq:chainruleeq] as \(x\) goes to \(c\) exists and is equal to \(f'\bigl(g(c)\bigr) g'(c)\). Thus \(h\) is differentiable at \(c\) and the limit is \(f'\bigl(g(c)\bigr)g'(c)\).

## Exercises

Prove the product rule. Hint: Use \(f(x) g(x) - f(c) g(c) = f(x)\bigl( g(x) - g(c) \bigr) + g(c) \bigl( f(x) - f(c) \bigr)\).

Prove the quotient rule. Hint: You can do this directly, but it may be easier to find the derivative of \(\nicefrac{1}{x}\) and then use the chain rule and the product rule.

[exercise:diffofxn] For \(n \in {\mathbb{Z}}\), prove that \(x^n\) is differentiable and find the derivative, unless, of course, \(n < 0\) and \(x=0\). Hint: Use the product rule.

Prove that a polynomial is differentiable and find the derivative. Hint: Use the previous exercise.

Define \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) by \[f(x) := \begin{cases} x^2 & \text{ if $x \in {\mathbb{Q}}$,}\\ 0 & \text{ otherwise.} \end{cases}\] Prove that \(f\) is differentiable at \(0\), but discontinuous at all points except \(0\).

Assume the inequality \(\left\lvert {x-\sin(x)} \right\rvert \leq x^2\). Prove that sin is differentiable at \(0\), and find the derivative at \(0\).

Using the previous exercise, prove that sin is differentiable at all \(x\) and that the derivative is \(\cos(x)\). Hint: Use the sum-to-product trigonometric identity as we did before.

Let \(f \colon I \to {\mathbb{R}}\) be differentiable. Given \(n \in {\mathbb{Z}}\), define \(f^n\) be the function defined by \(f^n(x) := {\bigl( f(x) \bigr)}^n\). If \(n < 0\) assume \(f(x) \not= 0\). Prove that \((f^n)'(x) = n {\bigl(f(x) \bigr)}^{n-1} f'(x)\).

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a differentiable Lipschitz continuous function. Prove that \(f'\) is a bounded function.

Let \(I_1, I_2\) be intervals. Let \(f \colon I_1 \to I_2\) be a bijective function and \(g \colon I_2 \to I_1\) be the inverse. Suppose that both \(f\) is differentiable at \(c \in I_1\) and \(f'(c) \not=0\) and \(g\) is differentiable at \(f(c)\). Use the chain rule to find a formula for \(g'\bigl(f(c)\bigr)\) (in terms of \(f'(c)\)).

[exercise:bndmuldiff] Suppose \(f \colon I \to {\mathbb{R}}\) is a bounded function and \(g \colon I \to {\mathbb{R}}\) is a function differentiable at \(c \in I\) and \(g(c) = g'(c) = 0\). Show that \(h(x) := f(x) g(x)\) is differentiable at \(c\). Hint: Note that you cannot apply the product rule.

[exercise:diffsqueeze] Suppose \(f \colon I \to {\mathbb{R}}\), \(g \colon I \to {\mathbb{R}}\), and \(h \colon I \to {\mathbb{R}}\), are functions. Suppose \(c \in I\) is such that \(f(c) = g(c) = h(c)\), \(g\) and \(h\) are differentiable at \(c\), and \(g'(c) = h'(c)\). Furthermore suppose \(h(x) \leq f(x) \leq g(x)\) for all \(x \in I\). Prove \(f\) is differentiable at \(c\) and \(f'(c) = g'(c) = h'(c)\).

### Mean value theorem

*Note: 2 lectures (some applications may be skipped)*

#### Relative minima and maxima

We talked about absolute maxima and minima. These are the tallest peaks and lowest valleys in the whole mountain range. We might also want to talk about peaks of individual mountains and valleys.

Let \(S \subset {\mathbb{R}}\) be a set and let \(f \colon S \to {\mathbb{R}}\) be a function. The function \(f\) is said to have a *relative maximum* at \(c \in S\) if there exists a \(\delta>0\) such that for all \(x \in S\) such that \(\left\lvert {x-c} \right\rvert < \delta\) we have \(f(x) \leq f(c)\). The definition of *relative minimum* is analogous.

[relminmax:thm] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a function differentiable at \(c \in (a,b)\), and \(c\) is a relative minimum or a relative maximum of \(f\). Then \(f'(c) = 0\).

We prove the statement for a maximum. For a minimum the statement follows by considering the function \(-f\).

Let \(c\) be a relative maximum of \(f\). In particular as long as \(\left\lvert {x-c} \right\rvert < \delta\) we have \(f(x)-f(c) \leq 0\). Then we look at the difference quotient. If \(x > c\) we note that \[\frac{f(x)-f(c)}{x-c} \leq 0 ,\] and if \(y < c\) we have \[\frac{f(y)-f(c)}{y-c} \geq 0 .\] See for an illustration.

We now take sequences \(\{ x_n\}\) and \(\{ y_n \}\), such that \(x_n > c\), and \(y_n < c\) for all \(n \in {\mathbb{N}}\), and such that \(\lim\, x_n = \lim\, y_n = c\). Since \(f\) is differentiable at \(c\) we know \[0 \geq \lim_{n\to\infty} \frac{f(x_n)-f(c)}{x_n-c} = f'(c) = \lim_{n\to\infty} \frac{f(y_n)-f(c)}{y_n-c} \geq 0. \qedhere\]

For a differentiable function, a point where \(f'(c) = 0\) is called a *critical point*. When \(f\) is not differentiable at some points, it is common to also say \(c\) is a critical point if \(f'(c)\) does not exist. The theorem says that a relative minimum or maximum at an interior point of an interval must be a critical point. As you remember from calculus, finding minima and maxima of a function can be done by finding all the critical points together with the endpoints of the interval and simply checking where is the function biggest or smallest.

#### Rolle’s theorem

Suppose a function is zero at both endpoints of an interval. Intuitively it ought to attain a minimum or a maximum in the interior of the interval, then at such a minimum or a maximum, the derivative should be zero. See for the geometric idea. This is the content of the so-called Rolle’s theorem ^{18}.

[thm:rolle] Let \(f \colon [a,b] \to {\mathbb{R}}\) be continuous function differentiable on \((a,b)\) such that \(f(a) = f(b)\). Then there exists a \(c \in (a,b)\) such that \(f'(c) = 0\).

As \(f\) is continuous on \([a,b]\) it attains an absolute minimum and an absolute maximum in \([a,b]\). We wish to apply and so we need a minimum or maximum at some \(c \in (a,b)\). Write \(K := f(a) = f(b)\). If there exists an \(x\) such that \(f(x) > K\), then the absolute maximum is bigger than \(K\) and hence occurs at \(c \in (a,b)\), and therefore we get \(f'(c) = 0\). On the other hand if there exists an \(x\) such that \(f(x) < K\), then the absolute minimum occurs at some \(c \in (a,b)\) and we have that \(f'(c) = 0\). If there is no \(x\) such that \(f(x) > K\) or \(f(x) < K\), then we have that \(f(x) = K\) for all \(x\) and then \(f'(x) = 0\) for all \(x \in [a,b]\), so any \(c\) will work.

Note that it is absolutely necessary for the derivative to exist for all \(x \in (a,b)\). For example take the function \(f(x) = \left\lvert {x} \right\rvert\) on \([-1,1]\). Clearly \(f(-1) = f(1)\), but there is no point where \(f'(c) = 0\).

#### Mean value theorem

We extend to functions that attain different values at the endpoints.

[thm:mvt] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function differentiable on \((a,b)\). Then there exists a point \(c \in (a,b)\) such that \[f(b)-f(a) = f'(c)(b-a) .\]

The theorem follows from . Define the function \(g \colon [a,b] \to {\mathbb{R}}\) by \[g(x) := f(x)-f(b)+\bigl(f(b)-f(a)\bigr)\frac{b-x}{b-a}.\] The function \(g\) is a differentiable on \((a,b)\), continuous on \([a,b]\), such that \(g(a) = 0\) and \(g(b) = 0\). Thus there exists \(c \in (a,b)\) such that \(g'(c) = 0\). \[0 = g'(c) = f'(c) + \bigl(f(b)-f(a)\bigr)\frac{-1}{b-a} .\] Or in other words \(f'(c)(b-a) = f(b)-f(a)\).

For a geometric interpretation of the mean value theorem, see . The idea is that the value \(\frac{f(b)-f(a)}{b-a}\) is the slope of the line between the points \(\bigl(a,f(a)\bigr)\) and \(\bigl(b,f(b)\bigr)\). Then \(c\) is the point such that \(f'(c) = \frac{f(b)-f(a)}{b-a}\), that is, the tangent line at the point \(\bigl(c,f(c)\bigr)\) has the same slope as the line between \(\bigl(a,f(a)\bigr)\) and \(\bigl(b,f(b)\bigr)\).

#### Applications

We now solve our very first differential equation.

[prop:derzeroconst] Let \(I\) be an interval and let \(f \colon I \to {\mathbb{R}}\) be a differentiable function such that \(f'(x) = 0\) for all \(x \in I\). Then \(f\) is constant.

Take arbitrary \(x,y \in I\) with \(x < y\). Then \(f\) restricted to \([x,y]\) satisfies the hypotheses of the . Therefore there is a \(c \in (x,y)\) such that \[f(y)-f(x) = f'(c)(y-x).\] as \(f'(c) = 0\), we have \(f(y) = f(x)\). Therefore, the function is constant.

Now that we know what it means for the function to stay constant, let us look at increasing and decreasing functions. We say \(f \colon I \to {\mathbb{R}}\) is *increasing* (resp. *strictly increasing*) if \(x < y\) implies \(f(x) \leq f(y)\) (resp. \(f(x) < f(y)\)). We define *decreasing* and *strictly decreasing* in the same way by switching the inequalities for \(f\).

[incdecdiffprop] Let \(I\) be an interval and let \(f \colon I \to {\mathbb{R}}\) be a differentiable function.

- \(f\) is increasing if and only if \(f'(x) \geq 0\) for all \(x \in I\).
- \(f\) is decreasing if and only if \(f'(x) \leq 0\) for all \(x \in I\).

Let us prove the first item. Suppose \(f\) is increasing, then for all \(x\) and \(c\) in \(I\) we have \[\frac{f(x)-f(c)}{x-c} \geq 0 .\] Taking a limit as \(x\) goes to \(c\) we see that \(f'(c) \geq 0\).

For the other direction, suppose \(f'(x) \geq 0\) for all \(x \in I\). Take any \(x, y \in I\) where \(x < y\). By the there is some \(c \in (x,y)\) such that \[f(x)-f(y) = f'(c)(x-y) .\] As \(f'(c) \geq 0\), and \(x-y < 0\), then \(f(x) - f(y) \leq 0\) or \(f(x) \leq f(y)\) and so \(f\) is increasing.

We leave the decreasing part to the reader as exercise.

We can make a similar but weaker statement about strictly increasing and decreasing functions. If \(f'(x) > 0\) for all \(x \in I\), then \(f\) is strictly increasing. The proof is left as an exercise. The converse is not true. For example, \(f(x) := x^3\) is a strictly increasing function, but \(f'(0) = 0\).

Another application of the is the following result about location of extrema. The theorem is stated for an absolute minimum and maximum, but the way it is applied to find relative minima and maxima is to restrict \(f\) to an interval \((c-\delta,c+\delta)\).

[firstderminmaxtest] Let \(f \colon (a,b) \to {\mathbb{R}}\) be continuous. Let \(c \in (a,b)\) and suppose \(f\) is differentiable on \((a,c)\) and \((c,b)\).

- If \(f'(x) \leq 0\) for \(x \in (a,c)\) and \(f'(x) \geq 0\) for \(x \in (c,b)\), then \(f\) has an absolute minimum at \(c\).
- If \(f'(x) \geq 0\) for \(x \in (a,c)\) and \(f'(x) \leq 0\) for \(x \in (c,b)\), then \(f\) has an absolute maximum at \(c\).

Let us prove the first item. The second is left to the reader. Let \(x\) be in \((a,c)\) and \(\{ y_n\}\) a sequence such that \(x < y_n < c\) and \(\lim\, y_n = c\). By the previous proposition, the function is decreasing on \((a,c)\) so \(f(x) \geq f(y_n)\). The function is continuous at \(c\) so we can take the limit to get \(f(x) \geq f(c)\) for all \(x \in (a,c)\).

Similarly take \(x \in (c,b)\) and \(\{ y_n\}\) a sequence such that \(c < y_n < x\) and \(\lim\, y_n = c\). The function is increasing on \((c,b)\) so \(f(x) \geq f(y_n)\). By continuity of \(f\) we get \(f(x) \geq f(c)\) for all \(x \in (c,b)\). Thus \(f(x) \geq f(c)\) for all \(x \in (a,b)\).

The converse of the proposition does not hold. See below.

#### Continuity of derivatives and the intermediate value theorem

Derivatives of functions satisfy an intermediate value property. The result is usually called Darboux’s theorem.

[thm:darboux] Let \(f \colon [a,b] \to {\mathbb{R}}\) be differentiable. Suppose that there exists a \(y \in {\mathbb{R}}\) such that \(f'(a) < y < f'(b)\) or \(f'(a) > y > f'(b)\). Then there exists a \(c \in (a,b)\) such that \(f'(c) = y\).

Suppose without loss of generality that \(f'(a) < y < f'(b)\). Define \[g(x) := yx - f(x) .\] As \(g\) is continuous on \([a,b]\), then \(g\) attains a maximum at some \(c \in [a,b]\).

Now compute \(g'(x) = y-f'(x)\). Thus \(g'(a) > 0\). As the derivative is the limit of difference quotients and is positive, there must be some difference quotient that is positive. That is, there must exist an \(x > a\) such that \[\frac{g(x)-g(a)}{x-a} > 0 ,\] or \(g(x) > g(a)\). Thus \(a\) cannot possibly be a maximum of \(g\). Similarly as \(g'(b) < 0\), we find an \(x < b\) (a different \(x\)) such that \(\frac{g(x)-g(b)}{x-b} < 0\) or that \(g(x) > g(b)\), thus \(b\) cannot possibly be a maximum.

Therefore \(c \in (a,b)\). Then as \(c\) is a maximum of \(g\) we find \(g'(c) = 0\) and \(f'(c) = y\).

We have seen already that there exist discontinuous functions that have the intermediate value property. While it is hard to imagine at first, there also exist functions that are differentiable everywhere and the derivative is not continuous.

[baddifffunc:example] Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be the function defined by \[f(x) := \begin{cases} {\bigl( x \sin(\nicefrac{1}{x}) \bigr)}^2 & \text{ if $x \not= 0$,} \\ 0 & \text{ if $x = 0$.} \end{cases}\] We claim that \(f\) is differentiable everywhere, but \(f' \colon {\mathbb{R}}\to {\mathbb{R}}\) is not continuous at the origin. Furthermore, \(f\) has a minimum at 0, but the derivative changes sign infinitely often near the origin. See .

Proof: It is easy to see from the definition that \(f\) has an absolute minimum at 0: we know \(f(x) \geq 0\) for all \(x\) and \(f(0) = 0\).

The function \(f\) is differentiable for \(x\not=0\) and the derivative is \(2 \sin (\nicefrac{1}{x}) \bigl( x \sin (\nicefrac{1}{x}) - \cos(\nicefrac{1}{x}) \bigr)\). As an exercise show that for \(x_n = \frac{4}{(8n+1)\pi}\) we have \(\lim\, f'(x_n) = -1\), and for \(y_n = \frac{4}{(8n+3)\pi}\) we have \(\lim\, f'(y_n) = 1\). Hence if \(f'\) exists at \(0\), then it cannot be continuous.

Let us show that \(f'\) exists at 0. We claim that the derivative is zero. In other words \(\left\lvert {\frac{f(x)-f(0)}{x-0} - 0} \right\rvert\) goes to zero as \(x\) goes to zero. For \(x \not= 0\) we have \[\left\lvert {\frac{f(x)-f(0)}{x-0} - 0} \right\rvert = \left\lvert {\frac{x^2 \sin^2(\nicefrac{1}{x})}{x}} \right\rvert = \left\lvert {x \sin^2(\nicefrac{1}{x})} \right\rvert \leq \left\lvert {x} \right\rvert .\] And, of course, as \(x\) tends to zero, then \(\left\lvert {x} \right\rvert\) tends to zero and hence \(\left\lvert {\frac{f(x)-f(0)}{x-0} - 0} \right\rvert\) goes to zero. Therefore, \(f\) is differentiable at 0 and the derivative at 0 is 0. A key point in the above calculation is that is that \(\left\lvert {f(x)} \right\rvert \leq x^2\), see also Exercises [exercise:bndmuldiff] and [exercise:diffsqueeze].

It is sometimes useful to assume the derivative of a differentiable function is continuous. If \(f \colon I \to {\mathbb{R}}\) is differentiable and the derivative \(f'\) is continuous on \(I\), then we say \(f\) is *continuously differentiable*. It is common to write \(C^1(I)\) for the set of continuously differentiable functions on \(I\).

## Exercises

Finish the proof of .

Finish the proof of .

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a differentiable function such that \(f'\) is a bounded function. Prove \(f\) is a Lipschitz continuous function.

Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is differentiable and \(c \in [a,b]\). Then show there exists a sequence \(\{ x_n \}\) converging to \(c\), \(x_n \not= c\) for all \(n\), such that \[f'(c) = \lim_{n\to \infty} f'(x_n).\] Do note this does *not* imply that \(f'\) is continuous (why?).

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a function such that \(\left\lvert {f(x)-f(y)} \right\rvert \leq \left\lvert {x-y} \right\rvert^2\) for all \(x\) and \(y\). Show that \(f(x) = C\) for some constant \(C\). Hint: Show that \(f\) is differentiable at all points and compute the derivative.

[exercise:posderincr] Suppose \(I\) is an interval and \(f \colon I \to {\mathbb{R}}\) is a differentiable function. If \(f'(x) > 0\) for all \(x \in I\), show that \(f\) is strictly increasing.

Suppose \(f \colon (a,b) \to {\mathbb{R}}\) is a differentiable function such that \(f'(x) \not= 0\) for all \(x \in (a,b)\). Suppose there exists a point \(c \in (a,b)\) such that \(f'(c) > 0\). Prove \(f'(x) > 0\) for all \(x \in (a,b)\).

[exercise:samediffconst] Suppose \(f \colon (a,b) \to {\mathbb{R}}\) and \(g \colon (a,b) \to {\mathbb{R}}\) are differentiable functions such that \(f'(x) = g'(x)\) for all \(x \in (a,b)\), then show that there exists a constant \(C\) such that \(f(x) = g(x) + C\).

Prove the following version of L’Hopital’s rule. Suppose \(f \colon (a,b) \to {\mathbb{R}}\) and \(g \colon (a,b) \to {\mathbb{R}}\) are differentiable functions. Suppose that at \(c \in (a,b)\), \(f(c) = 0\), \(g(c)=0\), and that the limit of \(\nicefrac{f'(x)}{g'(x)}\) as \(x\) goes to \(c\) exists. Show that \[\lim_{x \to c} \frac{f(x)}{g(x)} = \lim_{x \to c} \frac{f'(x)}{g'(x)} .\]

Let \(f \colon (a,b) \to {\mathbb{R}}\) be an unbounded differentiable function. Show \(f' \colon (a,b) \to {\mathbb{R}}\) is unbounded.

Prove the theorem Rolle actually proved in 1691: *If \(f\) is a polynomial, \(f'(a) = f'(b) = 0\) for some \(a < b\), and there is no \(c \in (a,b)\) such that \(f'(c) = 0\), then there is at most one root of \(f\) in \((a,b)\), that is at most one \(x \in (a,b)\) such that \(f(x) = 0\).* In other words, between any two consecutive roots of \(f'\) is at most one root of \(f\). Hint: suppose there are two roots and see what happens.

Suppose \(a,b \in {\mathbb{R}}\) and \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is differentiable, \(f'(x) = a\) for all \(x\), and \(f(0) = b\). Find \(f\) and prove that it is the unique differentiable function with this property.

### Taylor’s theorem

*Note: half a lecture (optional section)*

#### Derivatives of higher orders

When \(f \colon I \to {\mathbb{R}}\) is differentiable, we obtain a function \(f' \colon I \to {\mathbb{R}}\). The function \(f'\) is called the *first derivative* of \(f\). If \(f'\) is differentiable, we denote by \(f'' \colon I \to {\mathbb{R}}\) the derivative of \(f'\). The function \(f''\) is called the *second derivative* of \(f\). We similarly obtain \(f'''\), \(f''''\), and so on. With a larger number of derivatives the notation would get out of hand; we denote by \(f^{(n)}\) the *\(n\)th derivative* of \(f\).

When \(f\) possesses \(n\) derivatives, we say \(f\) is *\(n\) times differentiable*.

#### Taylor’s theorem

Taylor’s theorem ^{19} is a generalization of the . Mean value theorem says that up to a small error \(f(x)\) for \(x\) near \(x_0\) can be approximated by \(f(x_0)\), that is \[f(x) = f(x_0) + f'(c)(x-x_0),\] where the “error” is measured in terms of the first derivative at some point \(c\) between \(x\) and \(x_0\). Taylor’s theorem generalizes this result to higher derivatives. It tells us that up to a small error, any \(n\) times differentiable function can be approximated at a point \(x_0\) by a polynomial. The error of this approximation behaves like \({(x-x_0)}^{n}\) near the point \(x_0\). To see why this is a good approximation notice that for a big \(n\), \({(x-x_0)}^n\) is very small in a small interval around \(x_0\).

For an \(n\) times differentiable function \(f\) defined near a point \(x_0 \in {\mathbb{R}}\), define the \(n\)th *Taylor polynomial* for \(f\) at \(x_0\) as \[\begin{split} P_n^{x_0}(x) & := \sum_{k=0}^n \frac{f^{(k)}(x_0)}{k!}{(x-x_0)}^k \\ & = f(x_0) + f'(x_0)(x-x_0) + \frac{f''(x_0)}{2}{(x-x_0)}^2 + \frac{f^{(3)}(x_0)}{6}{(x-x_0)}^3 + \cdots + \frac{f^{(n)}(x_0)}{n!}{(x-x_0)}^n . \end{split}\]

Taylor’s theorem says a function behaves like its \(n\)th Taylor polynomial. The is really Taylor’s theorem for the first derivative.

[thm:taylor] Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is a function with \(n\) continuous derivatives on \([a,b]\) and such that \(f^{(n+1)}\) exists on \((a,b)\). Given distinct points \(x_0\) and \(x\) in \([a,b]\), we can find a point \(c\) between \(x_0\) and \(x\) such that \[f(x)=P_{n}^{x_0}(x)+\frac{f^{(n+1)}(c)}{(n+1)!}{(x-x_0)}^{n+1} .\]

The term \(R_n^{x_0}(x):=\frac{f^{(n+1)}(c)}{(n+1)!}{(x-x_0)}^{n+1}\) is called the *remainder term*. This form of the remainder term is called the *Lagrange form* of the remainder. There are other ways to write the remainder term, but we skip those. Note that \(c\) depends on both \(x\) and \(x_0\).

Find a number \(M_{x,x_0}\) (depending on \(x\) and \(x_0\)) solving the equation \[f(x)=P_{n}^{x_0}(x)+M_{x,x_0}{(x-x_0)}^{n+1} .\] Define a function \(g(s)\) by \[g(s) := f(s)-P_n^{x_0}(s)-M_{x,x_0}{(s-x_0)}^{n+1} .\] We compute the \(k\)th derivative at \(x_0\) of the Taylor polynomial \({(P_n^{x_0})}^{(k)}(x_0) = f^{(k)}(x_0)\) for \(k=0,1,2,\ldots,n\) (the zeroth derivative corresponds to the function itself). Therefore, \[g(x_0) = g'(x_0) = g''(x_0) = \cdots = g^{(n)}(x_0) = 0 .\] In particular \(g(x_0) = 0\). On the other hand \(g(x) = 0\). By the there exists an \(x_1\) between \(x_0\) and \(x\) such that \(g'(x_1) = 0\). Applying the to \(g'\) we obtain that there exists \(x_2\) between \(x_0\) and \(x_1\) (and therefore between \(x_0\) and \(x\)) such that \(g''(x_2) = 0\). We repeat the argument \(n+1\) times to obtain a number \(x_{n+1}\) between \(x_0\) and \(x_n\) (and therefore between \(x_0\) and \(x\)) such that \(g^{(n+1)}(x_{n+1}) = 0\).

Let \(c:=x_{n+1}\). We compute the \((n+1)\)th derivative of \(g\) to find \[g^{(n+1)}(s) = f^{(n+1)}(s)-(n+1)!\,M_{x,x_0} .\] Plugging in \(c\) for \(s\) we obtain \(M_{x,x_0} = \frac{f^{(n+1)}(c)}{(n+1)!}\), and we are done.

In the proof we have computed \({(P_n^{x_0})}^{(k)}(x_0) = f^{(k)}(x_0)\) for \(k=0,1,2,\ldots,n\). Therefore the Taylor polynomial has the same derivatives as \(f\) at \(x_0\) up to the \(n\)th derivative. That is why the Taylor polynomial is a good approximation to \(f\).

The definition of derivative says that a function is differentiable if it is locally approximated by a line. Similarly we mention in passing that there exists a converse to Taylor’s theorem, which we will neither state nor prove, saying that if a function is locally approximated in a certain way by a polynomial of degree \(d\), then it has \(d\) derivatives.

## Exercises

Compute the \(n\)th Taylor Polynomial at \(0\) for the exponential function.

Suppose \(p\) is a polynomial of degree \(d\). Given any \(x_0 \in {\mathbb{R}}\), show that the \((d+1)\)th Taylor polynomial for \(p\) at \(x_0\) is equal to \(p\).

Let \(f(x) := \left\lvert {x} \right\rvert^3\). Compute \(f'(x)\) and \(f''(x)\) for all \(x\), but show that \(f^{(3)}(0)\) does not exist.

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) has \(n\) continuous derivatives. Show that for any \(x_0 \in {\mathbb{R}}\), there exist polynomials \(P\) and \(Q\) of degree \(n\) and an \(\epsilon > 0\) such that \(P(x) \leq f(x) \leq Q(x)\) for all \(x \in [x_0-\epsilon,x_0+\epsilon]\) and \(Q(x)-P(x) = \lambda {(x-x_0)}^n\) for some \(\lambda \geq 0\).

If \(f \colon [a,b] \to {\mathbb{R}}\) has \(n+1\) continuous derivatives and \(x_0 \in [a,b]\), prove \(\lim\limits_{x\to x_0} \frac{R_n^{x_0}(x)}{{(x-x_0)}^n} = 0\).

Suppose \(f \colon [a,b] \to {\mathbb{R}}\) has \(n+1\) continuous derivatives and \(x_0 \in (a,b)\). Show that \(f^{(k)}(x_0) = 0\) for all \(k = 0, 1, 2, \ldots, n\) if and only if \(g(x) := \frac{f(x)}

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[26]/p[6]/span, line 1, column 8 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Suppose \(a,b,c \in {\mathbb{R}}\) and \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is differentiable, \(f''(x) = a\) for all \(x\), \(f'(0) = b\), and \(f(0) = c\). Find \(f\) and prove that it is the unique differentiable function with this property.

Show that a simple converse to Taylor’s theorem does not hold. Find a function \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) with no second derivative at \(x=0\) such that \(\left\lvert {f(x)} \right\rvert \leq \left\lvert {x^3} \right\rvert\), that is, \(f\) goes to zero at 0 faster than \(x^3\), and while \(f'(0)\) exists, \(f''(0)\) does not.

### Inverse function theorem

*Note: less than 1 lecture (optional section, needed for , requires )*

#### Inverse function theorem

The main idea of differentiating inverse functions is the following lemma.

[lemma:ift] Let \(I,J \subset {\mathbb{R}}\) be intervals. If \(f \colon I \to J\) is strictly monotone (hence one-to-one), onto (\(f(I) = J\)), differentiable at \(x\), and \(f'(x) \not= 0\), then the inverse \(f^{-1}\) is differentiable at \(y = f(x)\) and \[(f^{-1})'(y) = \frac{1}{f'\bigl( f^{-1}(y) \bigr)} = \frac{1}{f'(x)} .\] If \(f\) is continuously differentiable and \(f'\) is never zero, then \(f^{-1}\) is continuously differentiable.

By \(f\) has a continuous inverse. Let us call the inverse \(g \colon J \to I\) for convenience. Let \(x,y\) be as in the statement, take \(t \in I\) to be arbitrary and let \(s := f(t)\). Then \[\frac{g(s)-g(y)}{s-y} = \frac{g\bigl(f(t)\bigr)-g\bigl(f(x)\bigr)}{f(t)-f(x)} = \frac{t-x}{f(t)-f(x)} .\] As \(f\) is differentiable at \(x\) and \(f'(x) \not= 0\), then \(\frac{t-x}{f(t)-f(x)} \to \nicefrac{1}{f'(x)}\) as \(t \to x\). Because \(g(s) \to g(y)\) as \(s \to y\), we can plug in \(g(s)\) for \(t\), and \(g(y)\) for \(x\) and take the limit as \(s\) goes to \(y\), that is, the limit exists. In other words, \[\lim_{s \to y} \frac{g(s)-g(y)}{s-y} = \lim_{t \to x} \frac{t-x}{f(t)-f(x)} = \frac{1}{f'(x)} = \frac{1}{f'\bigl(g(y)\bigr)}\] See for the geometric idea.

If both \(f'\) and \(g\) are continuous, \(f'\) is nonzero at all \(x\), then the lemma applies at all points \(x \in I\) and the resulting function \(g'(y) = \frac{1}{f'\bigl(g(t)\bigr)}\) must be continuous.

What is usually called the inverse function theorem is the following result.

Let \(f \colon (a,b) \to {\mathbb{R}}\) be a continuously differentiable function, \(x_0 \in (a,b)\) a point where \(f'(x_0) \not= 0\). Then there exists an interval \(I \subset (a,b)\) with \(x_0 \in I\), the restriction \(f|_{I}\) is injective with an inverse \(g \colon J \to I\) defined on \(J := f(I)\), which is continuously differentiable and \[g'(y) = \frac{1}{f'\bigl( g(y) \bigr)} , \qquad \text{for all $y \in J$}.\]

Without loss of generality, suppose \(f'(x_0) > 0\). As \(f'\) is continuous, there must exist an interval \(I\) with \(x_0 \in I\) such that \(f'(x) > 0\) for all \(x_0 \in I\).

By \(f\) is strictly increasing on \(I\), and hence the restriction \(f|_{I}\) bijective onto \(J: = f(I)\). As \(f\) is continuous, then by the (see also ), \(f(I)\) is in interval. Now apply .

If you tried to prove the existence of roots directly as in you may have seen how difficult that endeavor is. However, with the machinery we have built for inverse functions it becomes an almost trivial exercise, and with and the inverse function theorem we prove far more than mere existence.

Given any \(n \in {\mathbb{N}}\) and any \(x \geq 0\) there exists a unique number \(y \geq 0\) (denoted \(x^{1/n} := y\)), such that \(y^n = x\). Furthermore, the function \(g \colon (0,\infty) \to (0,\infty)\) defined by \(g(x) := x^{1/n}\) is continuously differentiable and \[g'(x) = \frac{1}{nx^{(n-1)/n}} = \frac{1}{n} \, x^{(1-n)/n} ,\] using the convention \(x^{n/m} := {(x^{1/m})}^{n}\).

For \(x=0\) the existence of a unique root is trivial.

Let \(f(x) := x^n\). Using product rule, \(f\) is continuously differentiable and \(f'(x) = nx^{n-1}\), see . For \(x > 0\) the derivative \(f'\) is strictly positive and so again by , \(f\) is strictly increasing (this can also be proved directly). It is also easy to see that the image of \(f\) is the entire interval \((0,\infty)\). We obtain a unique inverse \(g\) and so the existence and uniqueness of positive \(n\)th roots. We apply to obtain the derivative.

The corollary provides a good example of where the inverse function theorem gives us an interval smaller than \((a,b)\). Take \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) defined by \(f(x) := x^2\). Then \(f'(x) \not= 0\) as long as \(x \not= 0\). If \(x_0 > 0\), we can take \(I=(0,\infty)\), but no larger.

Another useful example is \(f(x) := x^3\). The function \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is one-to-one and onto, so \(f^{-1}(x) = x^{1/3}\) exists on the entire real line including zero and negative \(x\). The function \(f\) has a continuous derivative, but \(f^{-1}\) has no derivative at the origin. The point is that \(f'(0) = 0\). See also .

## Exercises

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is continuously differentiable such that \(f'(x) > 0\) for all \(x\). Show that \(f\) is invertible on the interval \(J = f({\mathbb{R}})\), the inverse is continuously differentiable, and \({(f^{-1})}'(y) > 0\) for all \(y \in f({\mathbb{R}})\).

Suppose \(I,J\) are intervals and a monotone onto \(f \colon I \to J\) has an inverse \(g \colon J \to I\). Suppose you already know that both \(f\) and \(g\) are differentiable everywhere and \(f'\) is never zero. Using chain rule but not prove the formula \(g'(y) = \nicefrac{1}{f'\bigl(g(y)\bigr)}\).

Let \(n\in {\mathbb{N}}\) be even. Prove that every \(x > 0\) has a unique negative \(n\)th root. That is, there exists a negative number \(y\) such that \(y^n = x\). Compute the derivative of the function \(g(x) := y\).

[exercise:oddroot] Let \(n \in {\mathbb{N}}\) be odd and \(n \geq 3\). Prove that every \(x\) has a unique \(n\)th root. That is, there exists a number \(y\) such that \(y^n = x\). Prove that the function defined by \(g(x) := y\) is differentiable except at \(x=0\) and compute the derivative. Prove that \(g\) is not differentiable at \(x=0\).

§3] Show that if in the inverse function theorem \(f\) has \(k\) continuous derivatives, then the inverse function \(g\) also has \(k\) continuous derivatives.

Let \(f(x) := x + 2 x^2 \sin(\nicefrac{1}{x})\) for \(x \not= 0\) and \(f(0) = 0\). Show that \(f\) is differentiable at all \(x\), that \(f'(0) > 0\), but that \(f\) is not invertible on any interval containing the origin.

a) Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be a continuously differentiable function and \(k > 0\) be a number such that \(f'(x) \geq k\) for all \(x \in {\mathbb{R}}\). Show \(f\) is one-to-one and onto, and has a continuously differentiable inverse \(f^{-1} \colon {\mathbb{R}}\to {\mathbb{R}}\). b) Find an example \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) where \(f'(x) > 0\) for all \(x\), but \(f\) is not onto.

Suppose \(I,J\) are intervals and a monotone onto \(f \colon I \to J\) has an inverse \(g \colon J \to I\). Suppose \(x \in I\) and \(y := f(x) \in J\), and that \(g\) is differentiable at \(y\). Prove:

a) If \(g'(y) \not= 0\), then \(f\) is differentiable at \(x\).

b) If \(g'(y) = 0\), then \(f\) is not differentiable at \(x\).

## The Riemann Integral

### The Riemann integral

*Note: 1.5 lectures*

We now get to the fundamental concept of integration. There is often confusion among students of calculus between *integral* and *antiderivative*. The integral is (informally) the area under the curve, nothing else. That we can compute an antiderivative using the integral is a nontrivial result we have to prove. In this chapter we define the *Riemann integral* ^{20} using the Darboux integral ^{21}, which is technically simpler than (but equivalent to) the traditional definition as done by Riemann.

#### Partitions and lower and upper integrals

We want to integrate a bounded function defined on an interval \([a,b]\). We first define two auxiliary integrals that can be defined for all bounded functions. Only then can we talk about the Riemann integral and the Riemann integrable functions.

A *partition* \(P\) of the interval \([a,b]\) is a finite set of numbers \(\{ x_0,x_1,x_2,\ldots,x_n \}\) such that \[a = x_0 < x_1 < x_2 < \cdots < x_{n-1} < x_n = b .\] We write \[\Delta x_i := x_i - x_{i-1} .\]

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function. Let \(P\) be a partition of \([a,b]\). Define \[\begin{aligned} & m_i := \inf \{ f(x) : x_{i-1} \leq x \leq x_i \} , \\ & M_i := \sup \{ f(x) : x_{i-1} \leq x \leq x_i \} , \\ & L(P,f) := \sum_{i=1}^n m_i \Delta x_i , \\ & U(P,f) := \sum_{i=1}^n M_i \Delta x_i .\end{aligned}\] We call \(L(P,f)\) the *lower Darboux sum* and \(U(P,f)\) the *upper Darboux sum*.

The geometric idea of Darboux sums is indicated in . The lower sum is the area of the shaded rectangles, and the upper sum is the area of the entire rectangles, shaded plus unshaded parts. The width of the \(i\)th rectangle is \(\Delta x_i\), the height of the shaded rectangle is \(m_i\) and the height of the entire rectangle is \(M_i\).

[sumulbound:prop] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function. Let \(m, M \in {\mathbb{R}}\) be such that for all \(x\) we have \(m \leq f(x) \leq M\). For any partition \(P\) of \([a,b]\) we have \[\label{sumulbound:eq} m(b-a) \leq L(P,f) \leq U(P,f) \leq M(b-a) .\]

Let \(P\) be a partition. Then note that \(m \leq m_i\) for all \(i\) and \(M_i \leq M\) for all \(i\). Also \(m_i \leq M_i\) for all \(i\). Finally \(\sum_{i=1}^n \Delta x_i = (b-a)\). Therefore, \[\begin{gathered} m(b-a) = m \left( \sum_{i=1}^n \Delta x_i \right) = \sum_{i=1}^n m \Delta x_i \leq \sum_{i=1}^n m_i \Delta x_i \leq \\ \leq \sum_{i=1}^n M_i \Delta x_i \leq \sum_{i=1}^n M \Delta x_i = M \left( \sum_{i=1}^n \Delta x_i \right) = M(b-a) .\end{gathered}\] Hence we get [sumulbound:eq]. In other words, the set of lower and upper sums are bounded sets.

As the sets of lower and upper Darboux sums are bounded, we define \[\begin{aligned} & \underline{\int_a^b} f(x)~dx := \sup \{ L(P,f) : P \text{ a partition of $[a,b]$} \} , \\ & \overline{\int_a^b} f(x)~dx := \inf \{ U(P,f) : P \text{ a partition of $[a,b]$} \} .\end{aligned}\] We call \(\underline{\int}\) the *lower Darboux integral* and \(\overline{\int}\) the *upper Darboux integral*. To avoid worrying about the variable of integration, we often simply write \[\underline{\int_a^b} f := \underline{\int_a^b} f(x)~dx \qquad \text{and} \qquad \overline{\int_a^b} f := \overline{\int_a^b} f(x)~dx .\]

If integration is to make sense, then the lower and upper Darboux integrals should be the same number, as we want a single number to call *the integral*. However, these two integrals may in fact differ for some functions.

Take the Dirichlet function \(f \colon [0,1] \to {\mathbb{R}}\), where \(f(x) := 1\) if \(x \in {\mathbb{Q}}\) and \(f(x) := 0\) if \(x \notin {\mathbb{Q}}\). Then \[\underline{\int_0^1} f = 0 \qquad \text{and} \qquad \overline{\int_0^1} f = 1 .\] The reason is that for every \(i\) we have \(m_i = \inf \{ f(x) : x \in [x_{i-1},x_i] \} = 0\) and \(M_i = \sup \{ f(x) : x \in [x_{i-1},x_i] \} = 1\). Thus \[\begin{aligned} & L(P,f) = \sum_{i=1}^n 0 \cdot \Delta x_i = 0 , \\ & U(P,f) = \sum_{i=1}^n 1 \cdot \Delta x_i = \sum_{i=1}^n \Delta x_i = 1 .\end{aligned}\]

The same definition of \(\underline{\int_a^b} f\) and \(\overline{\int_a^b} f\) is used when \(f\) is defined on a larger set \(S\) such that \([a,b] \subset S\). In that case, we use the restriction of \(f\) to \([a,b]\) and we must ensure that the restriction is bounded on \([a,b]\).

To compute the integral we often take a partition \(P\) and make it finer. That is, we cut intervals in the partition into yet smaller pieces.

Let \(P = \{ x_0, x_1, \ldots, x_n \}\) and \(\widetilde{P} = \{ \widetilde{x}_0, \widetilde{x}_1, \ldots, \widetilde{x}_m \}\) be partitions of \([a,b]\). We say \(\widetilde{P}\) is a *refinement* of \(P\) if as sets \(P \subset \widetilde{P}\).

That is, \(\widetilde{P}\) is a refinement of a partition if it contains all the numbers in \(P\) and perhaps some other numbers in between. For example, \(\{ 0, 0.5, 1, 2 \}\) is a partition of \([0,2]\) and \(\{ 0, 0.2, 0.5, 1, 1.5, 1.75, 2 \}\) is a refinement. The main reason for introducing refinements is the following proposition.

[prop:refinement] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function, and let \(P\) be a partition of \([a,b]\). Let \(\widetilde{P}\) be a refinement of \(P\). Then \[L(P,f) \leq L(\widetilde{P},f) \qquad \text{and} \qquad U(\widetilde{P},f) \leq U(P,f) .\]

The tricky part of this proof is to get the notation correct. Let \(\widetilde{P} := \{ \widetilde{x}_0, \widetilde{x}_1, \ldots, \widetilde{x}_m \}\) be a refinement of \(P := \{ x_0, x_1, \ldots, x_n \}\). Then \(x_0 = \widetilde{x}_0\) and \(x_n = \widetilde{x}_m\). In fact, we can find integers \(k_0 < k_1 < \cdots < k_n\) such that \(x_j = \widetilde{x}_{k_j}\) for \(j=0,1,2,\ldots,n\).

Let \(\Delta \widetilde{x}_j = \widetilde{x}_{j-1} - \widetilde{x}_j\). We get \[\Delta x_j = \sum_{p=k_{j-1}+1}^{k_j} \Delta \widetilde{x}_p .\]

Let \(m_j\) be as before and correspond to the partition \(P\). Let \(\widetilde{m}_j := \inf \{ f(x) : \widetilde{x}_{j-1} \leq x \leq \widetilde{x}_j \}\). Now, \(m_j \leq \widetilde{m}_p\) for \(k_{j-1} < p \leq k_j\). Therefore, \[m_j \Delta x_j = m_j \sum_{p=k_{j-1}+1}^{k_j} \Delta \widetilde{x}_p = \sum_{p=k_{j-1}+1}^{k_j} m_j \Delta \widetilde{x}_p \leq \sum_{p=k_{j-1}+1}^{k_j} \widetilde{m}_p \Delta \widetilde{x}_p .\] So \[L(P,f) = \sum_{j=1}^n m_j \Delta x_j \leq \sum_{j=1}^n \sum_{p=k_{j-1}+1}^{k_j} \widetilde{m}_p \Delta \widetilde{x}_p = \sum_{j=1}^m \widetilde{m}_j \Delta \widetilde{x}_j = L(\widetilde{P},f).\]

The proof of \(U(\widetilde{P},f) \leq U(P,f)\) is left as an exercise.

Armed with refinements we prove the following. The key point of this next proposition is that the lower Darboux integral is less than or equal to the upper Darboux integral.

[intulbound:prop] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function. Let \(m, M \in {\mathbb{R}}\) be such that for all \(x \in [a,b]\) we have \(m \leq f(x) \leq M\). Then \[\label{intulbound:eq} m(b-a) \leq \underline{\int_a^b} f \leq \overline{\int_a^b} f \leq M(b-a) .\]

By we have for any partition \(P\) \[m(b-a) \leq L(P,f) \leq U(P,f) \leq M(b-a).\] The inequality \(m(b-a) \leq L(P,f)\) implies \(m(b-a) \leq \underline{\int_a^b} f\). Also \(U(P,f) \leq M(b-a)\) implies \(\overline{\int_a^b} f \leq M(b-a)\).

The key point of this proposition is the middle inequality in [intulbound:eq]. Let \(P_1, P_2\) be partitions of \([a,b]\). Define \(\widetilde{P} := P_1 \cup P_2\). The set \(\widetilde{P}\) is a partition of \([a,b]\). Furthermore, \(\widetilde{P}\) is a refinement of \(P_1\) and it is also a refinement of \(P_2\). By we have \(L(P_1,f) \leq L(\widetilde{P},f)\) and \(U(\widetilde{P},f) \leq U(P_2,f)\). Putting it all together we have \[L(P_1,f) \leq L(\widetilde{P},f) \leq U(\widetilde{P},f) \leq U(P_2,f) .\] In other words, for two arbitrary partitions \(P_1\) and \(P_2\) we have \(L(P_1,f) \leq U(P_2,f)\). Now we recall . Taking the supremum and infimum over all partitions we get \[\sup \{ L(P,f) : \text{$P$ a partition of $[a,b]$} \} \leq \inf \{ U(P,f) : \text{$P$ a partition of $[a,b]$} \} .\] In other words \(\underline{\int_a^b} f \leq \overline{\int_a^b} f\).

#### Riemann integral

We can finally define the Riemann integral. However, the Riemann integral is only defined on a certain class of functions, called the Riemann integrable functions.

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function such that \[\underline{\int_a^b} f(x)~dx = \overline{\int_a^b} f(x)~dx .\] Then \(f\) is said to be *Riemann integrable*. The set of Riemann integrable functions on \([a,b]\) is denoted by \({\mathcal{R}}[a,b]\). When \(f \in {\mathcal{R}}[a,b]\) we define \[\int_a^b f(x)~dx := \underline{\int_a^b} f(x)~dx = \overline{\int_a^b} f(x)~dx .\] As before, we often simply write \[\int_a^b f := \int_a^b f(x)~dx.\] The number \(\int_a^b f\) is called the *Riemann integral* of \(f\), or sometimes simply the *integral* of \(f\).

By definition, any Riemann integrable function is bounded. By appealing to we immediately obtain the following proposition.

[intbound:prop] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a Riemann integrable function. Let \(m, M \in {\mathbb{R}}\) be such that \(m \leq f(x) \leq M\) for all \(x \in [a,b]\). Then \[m(b-a) \leq \int_a^b f \leq M(b-a) .\]

Often we use a weaker form of this proposition. That is, if \(\left\lvert {f(x)} \right\rvert \leq M\) for all \(x \in [a,b]\), then \[\left\lvert {\int_a^b f} \right\rvert \leq M(b-a) .\]

We integrate constant functions using . If \(f(x) := c\) for some constant \(c\), then we take \(m = M = c\). In inequality [intulbound:eq] all the inequalities must be equalities. Thus \(f\) is integrable on \([a,b]\) and \(\int_a^b f = c(b-a)\).

Let \(f \colon [0,2] \to {\mathbb{R}}\) be defined by \[f(x) := \begin{cases} 1 & \text{ if $x < 1$,}\\ \nicefrac{1}{2} & \text{ if $x = 1$,}\\ 0 & \text{ if $x > 1$.} \end{cases}\] We claim \(f\) is Riemann integrable and \(\int_0^2 f = 1\).

Proof: Let \(0 < \epsilon < 1\) be arbitrary. Let \(P := \{0, 1-\epsilon, 1+\epsilon, 2\}\) be a partition. We use the notation from the definition of the Darboux sums. Then \[\begin{aligned} m_1 &= \inf \{ f(x) : x \in [0,1-\epsilon] \} = 1 , & M_1 &= \sup \{ f(x) : x \in [0,1-\epsilon] \} = 1 , \\ m_2 &= \inf \{ f(x) : x \in [1-\epsilon,1+\epsilon] \} = 0 , & M_2 &= \sup \{ f(x) : x \in [1-\epsilon,1+\epsilon] \} = 1 , \\ m_3 &= \inf \{ f(x) : x \in [1+\epsilon,2] \} = 0 , & M_3 &= \sup \{ f(x) : x \in [1+\epsilon,2] \} = 0 .\end{aligned}\] Furthermore, \(\Delta x_1 = 1-\epsilon\), \(\Delta x_2 = 2\epsilon\) and \(\Delta x_3 = 1-\epsilon\). We compute \[\begin{aligned} & L(P,f) = \sum_{i=1}^3 m_i \Delta x_i = 1 \cdot (1-\epsilon) + 0 \cdot 2\epsilon + 0 \cdot (1-\epsilon) = 1-\epsilon , \\ & U(P,f) = \sum_{i=1}^3 M_i \Delta x_i = 1 \cdot (1-\epsilon) + 1 \cdot 2\epsilon + 0 \cdot (1-\epsilon) = 1+\epsilon .\end{aligned}\] Thus, \[\overline{\int_0^2} f - \underline{\int_0^2} f \leq U(P,f) - L(P,f) = (1+\epsilon) - (1-\epsilon) = 2 \epsilon .\] By we have \(\underline{\int_0^2} f \leq \overline{\int_0^2} f\). As \(\epsilon\) was arbitrary we see \(\overline{\int_0^2} f = \underline{\int_0^2} f\). So \(f\) is Riemann integrable. Finally, \[1-\epsilon = L(P,f) \leq \int_0^2 f \leq U(P,f) = 1+\epsilon.\] Hence, \(\bigl\lvert \int_0^2 f - 1 \bigr\rvert \leq \epsilon\). As \(\epsilon\) was arbitrary, we have \(\int_0^2 f = 1\).

It may be worthwhile to extract part of the technique of the example into a proposition.

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function. Then \(f\) is Riemann integrable if for every \(\epsilon > 0\), there exists a partition \(P\) such that \[U(P,f) - L(P,f) < \epsilon .\]

If for every \(\epsilon > 0\), a \(P\) exists we have: \[0 \leq \overline{\int_a^b} f - \underline{\int_a^b} f \leq U(P,f) - L(P,f) < \epsilon .\] Therefore, \(\overline{\int_a^b} f = \underline{\int_a^b} f\), and \(f\) is integrable.

Let us show \(\frac{1}{1+x}\) is integrable on \([0,b]\) for any \(b > 0\). We will see later that all continuous functions are integrable, but let us demonstrate how we do it directly.

Let \(\epsilon > 0\) be given. Take \(n \in {\mathbb{N}}\) and pick \(x_j := \nicefrac{jb}{n}\), to form the partition \(P := \{ x_0,x_1,\ldots,x_n \}\) of \([0,b]\). We have \(\Delta x_j = \nicefrac{b}{n}\) for all \(j\). As \(f\) is decreasing, for any subinterval \([x_{j-1},x_j]\) we obtain \[m_j = \inf \left\{ \frac{1}{1+x} : x \in [x_{j-1},x_j] \right\} = \frac{1}{1+x_j} , \qquad M_j = \sup \left\{ \frac{1}{1+x} : x \in [x_{j-1},x_j] \right\} = \frac{1}{1+x_{j-1}} .\] Then we have \[\begin{gathered} U(P,f)-L(P,f) = \sum_{j=1}^n \Delta x_j (M_j-m_j) = \\ = \frac{b}{n} \sum_{j=1}^n \left( \frac{1}{1+\nicefrac{(j-1)b}{n}} - \frac{1}{1+\nicefrac{jb}{n}} \right) = \frac{b}{n} \left( \frac{1}{1+\nicefrac{0b}{n}} - \frac{1}{1+\nicefrac{nb}{n}} \right) = \frac{b^2}{n(b+1)} .\end{gathered}\] The sum telescopes, the terms successively cancel each other, something we have seen before. Picking \(n\) to be such that \(\frac{b^2}{n(b+1)} < \epsilon\) the proposition is satisfied and the function is integrable.

#### More notation

When \(f \colon S \to {\mathbb{R}}\) is defined on a larger set \(S\) and \([a,b] \subset S\), we say \(f\) is Riemann integrable on \([a,b]\) if the restriction of \(f\) to \([a,b]\) is Riemann integrable. In this case, we say \(f \in {\mathcal{R}}[a,b]\), and we write \(\int_a^b f\) to mean the Riemann integral of the restriction of \(f\) to \([a,b]\).

It is useful to define the integral \(\int_a^b f\) even if \(a \not< b\). Suppose \(b < a\) and \(f \in {\mathcal{R}}[b,a]\), then define \[\int_a^b f := - \int_b^a f .\] For any function \(f\) we define \[\int_a^a f := 0 .\]

At times, the variable \(x\) may already have some other meaning. When we need to write down the variable of integration, we may simply use a different letter. For example, \[\int_a^b f(s)~ds := \int_a^b f(x)~dx .\]

## Exercises

Let \(f \colon [0,1] \to {\mathbb{R}}\) be defined by \(f(x) := x^3\) and let \(P := \{ 0, 0.1, 0.4, 1 \}\). Compute \(L(P,f)\) and \(U(P,f)\).

Let \(f \colon [0,1] \to {\mathbb{R}}\) be defined by \(f(x) := x\). Show that \(f \in {\mathcal{R}}[0,1]\) and compute \(\int_0^1 f\) using the definition of the integral (but feel free to use the propositions of this section).

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function. Suppose there exists a sequence of partitions \(\{ P_k \}\) of \([a,b]\) such that \[\lim_{k \to \infty} \bigl( U(P_k,f) - L(P_k,f) \bigr) = 0 .\] Show that \(f\) is Riemann integrable and that \[\int_a^b f = \lim_{k \to \infty} U(P_k,f) = \lim_{k \to \infty} L(P_k,f) .\]

Finish the proof of .

Suppose \(f \colon [-1,1] \to {\mathbb{R}}\) is defined as \[f(x) := \begin{cases} 1 & \text{ if $x > 0$,} \\ 0 & \text{ if $x \leq 0$.} \end{cases}\] Prove that \(f \in {\mathcal{R}}[-1,1]\) and compute \(\int_{-1}^1 f\) using the definition of the integral (but feel free to use the propositions of this section).

Let \(c \in (a,b)\) and let \(d \in {\mathbb{R}}\). Define \(f \colon [a,b] \to {\mathbb{R}}\) as \[f(x) := \begin{cases} d & \text{ if $x = c$,} \\ 0 & \text{ if $x \not= c$.} \end{cases}\] Prove that \(f \in {\mathcal{R}}[a,b]\) and compute \(\int_a^b f\) using the definition of the integral (but feel free to use the propositions of this section).

[exercise:taggedpartition] Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is Riemann integrable. Let \(\epsilon > 0\) be given. Then show that there exists a partition \(P = \{ x_0, x_1, \ldots, x_n \}\) such that if we pick any set of numbers \(\{ c_1, c_2, \ldots, c_n \}\) with \(c_k \in [x_{k-1},x_k]\) for all \(k\), then \[\left\lvert {\int_a^b f - \sum_{k=1}^n f(c_k) \Delta x_k} \right\rvert < \epsilon .\]

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a Riemann integrable function. Let \(\alpha > 0\) and \(\beta \in {\mathbb{R}}\). Then define \(g(x) := f(\alpha x + \beta)\) on the interval \(I = [\frac{a-\beta}{\alpha}, \frac{b-\beta}{\alpha}]\). Show that \(g\) is Riemann integrable on \(I\).

Suppose \(f \colon [0,1] \to {\mathbb{R}}\) and \(g \colon [0,1] \to {\mathbb{R}}\) are such that for all \(x \in (0,1]\) we have \(f(x) = g(x)\). Suppose \(f\) is Riemann integrable. Prove \(g\) is Riemann integrable and \(\int_{0}^1 f = \int_{0}^1 g\).

Let \(f \colon [0,1] \to {\mathbb{R}}\) be a bounded function. Let \(P_n = \{ x_0,x_1,\ldots,x_n \}\) be a uniform partition of \([0,1]\), that is, \(x_j := \nicefrac{j}{n}\). Is \(\{ L(P_n,f) \}_{n=1}^\infty\) always monotone? Yes/No: Prove or find a counterexample.

For a bounded function \(f \colon [0,1] \to {\mathbb{R}}\) let \(R_n := (\nicefrac{1}{n})\sum_{j=1}^n f(\nicefrac{j}{n})\) (the uniform right hand rule). a) If \(f\) is Riemann integrable show \(\int_0^1 f = \lim \, R_n\). b) Find an \(f\) that is not Riemann integrable, but \(\lim \, R_n\) exists.

[exercise:riemannintdarboux] Generalize the previous exercise. Show that \(f \in {\mathcal{R}}[a,b]\) if and only if there exists an \(I \in {\mathbb{R}}\), such that for every \(\epsilon > 0\) there exists a \(\delta > 0\) such that if \(P\) is a partition with \(\Delta x_i < \delta\) for all \(i\), then \(\left\lvert {L(P,f) - I} \right\rvert < \epsilon\) and \(\left\lvert {U(P,f) - I} \right\rvert < \epsilon\). If \(f \in {\mathcal{R}}[a,b]\), then \(I = \int_a^b f\).

[exercise:riemannintdarboux2] Using and the idea of the proof in , show that Darboux integral is the same as the standard definition of Riemann integral, which you have most likely seen in calculus. That is, show that \(f \in {\mathcal{R}}[a,b]\) if and only if there exists an \(I \in {\mathbb{R}}\), such that for every \(\epsilon > 0\) there exists a \(\delta > 0\) such that if \(P = \{ x_0,x_1,\ldots,x_n \}\) is a partition with \(\Delta x_i < \delta\) for all \(i\), then \(\left\lvert {\sum_{i=1}^n f(c_i) \Delta x_i - I} \right\rvert < \epsilon\) for any set \(\{ c_1,c_2,\ldots,c_n \}\) with \(c_i \in [x_{i-1},x_i]\). If \(f \in {\mathcal{R}}[a,b]\), then \(I = \int_a^b f\).

Find an example of functions \(f \colon [0,1] \to {\mathbb{R}}\) which is Riemann integrable, and \(g \colon [0,1] \to [0,1]\) which is one-to-one and onto, such that the composition \(f \circ g\) is not Riemann integrable.

### Properties of the integral

*Note: 2 lectures, integrability of functions with discontinuities can safely be skipped*

#### Additivity

The next result we prove is usually referred to as the additive property of the integral. First we prove the additivity property for the lower and upper Darboux integrals.

[lemma:darbouxadd] Suppose \(a < b < c\) and \(f \colon [a,c] \to {\mathbb{R}}\) is a bounded function. Then \[\underline{\int_a^c} f = \underline{\int_a^b} f + \underline{\int_b^c} f\] and \[\overline{\int_a^c} f = \overline{\int_a^b} f + \overline{\int_b^c} f .\]

If we have partitions \(P_1 = \{ x_0,x_1,\ldots,x_k \}\) of \([a,b]\) and \(P_2 = \{ x_k, x_{k+1}, \ldots, x_n \}\) of \([b,c]\), then the set \(P := P_1 \cup P_2 = \{ x_0, x_1, \ldots, x_n \}\) is a partition of \([a,c]\). Then \[L(P,f) = \sum_{j=1}^n m_j \Delta x_j = \sum_{j=1}^k m_j \Delta x_j + \sum_{j=k+1}^n m_j \Delta x_j = L(P_1,f) + L(P_2,f) .\] When we take the supremum of the right hand side over all \(P_1\) and \(P_2\), we are taking a supremum of the left hand side over all partitions \(P\) of \([a,c]\) that contain \(b\). If \(Q\) is any partition of \([a,c]\) and \(P = Q \cup \{ b \}\), then \(P\) is a refinement of \(Q\) and so \(L(Q,f) \leq L(P,f)\). Therefore, taking a supremum only over the \(P\) that contain \(b\) is sufficient to find the supremum of \(L(P,f)\) over all partitions \(P\), see . Finally recall to compute \[\begin{split} \underline{\int_a^c} f & = \sup \{ L(P,f) : \text{$P$ a partition of $[a,c]$} \} \\ & = \sup \{ L(P,f) : \text{$P$ a partition of $[a,c]$, $b \in P$} \} \\ & = \sup \{ L(P_1,f) + L(P_2,f) : \text{$P_1$ a partition of $[a,b]$, $P_2$ a partition of $[b,c]$} \} \\ & = \sup \{ L(P_1,f) : \text{$P_1$ a partition of $[a,b]$} \} + \sup \{ L(P_2,f) : \text{$P_2$ a partition of $[b,c]$} \} \\ &= \underline{\int_a^b} f + \underline{\int_b^c} f . \end{split}\]

Similarly, for \(P\), \(P_1\), and \(P_2\) as above we obtain \[U(P,f) = \sum_{j=1}^n M_j \Delta x_j = \sum_{j=1}^k M_j \Delta x_j + \sum_{j=k+1}^n M_j \Delta x_j = U(P_1,f) + U(P_2,f) .\] We wish to take the infimum on the right over all \(P_1\) and \(P_2\), and so we are taking the infimum over all partitions \(P\) of \([a,c]\) that contain \(b\). If \(Q\) is any partition of \([a,c]\) and \(P = Q \cup \{ b \}\), then \(P\) is a refinement of \(Q\) and so \(U(Q,f) \geq U(P,f)\). Therefore, taking an infimum only over the \(P\) that contain \(b\) is sufficient to find the infimum of \(U(P,f)\) for all \(P\). We obtain \[\overline{\int_a^c} f = \overline{\int_a^b} f + \overline{\int_b^c} f . \qedhere\]

Let \(a < b < c\). A function \(f \colon [a,c] \to {\mathbb{R}}\) is Riemann integrable if and only if \(f\) is Riemann integrable on \([a,b]\) and \([b,c]\). If \(f\) is Riemann integrable, then \[\int_a^c f = \int_a^b f + \int_b^c f .\]

Suppose \(f \in {\mathcal{R}}[a,c]\), then \(\overline{\int_a^c} f = \underline{\int_a^c} f = \int_a^c f\). We apply the lemma to get \[\int_a^c f = \underline{\int_a^c} f = \underline{\int_a^b} f + \underline{\int_b^c} f \leq \overline{\int_a^b} f + \overline{\int_b^c} f = \overline{\int_a^c} f = \int_a^c f .\] Thus the inequality is an equality and \[\underline{\int_a^b} f + \underline{\int_b^c} f = \overline{\int_a^b} f + \overline{\int_b^c} f .\] As we also know \(\underline{\int_a^b} f \leq \overline{\int_a^b} f\) and \(\underline{\int_b^c} f \leq \overline{\int_b^c} f\), we conclude \[\underline{\int_a^b} f = \overline{\int_a^b} f \qquad \text{and} \qquad \underline{\int_b^c} f = \overline{\int_b^c} f .\] Thus \(f\) is Riemann integrable on \([a,b]\) and \([b,c]\) and the desired formula holds.

Now assume the restrictions of \(f\) to \([a,b]\) and to \([b,c]\) are Riemann integrable. We again apply the lemma to get \[\underline{\int_a^c} f = \underline{\int_a^b} f + \underline{\int_b^c} f = \int_a^b f + \int_b^c f = \overline{\int_a^b} f + \overline{\int_b^c} f = \overline{\int_a^c} f .\] Therefore \(f\) is Riemann integrable on \([a,c]\), and the integral is computed as indicated.

An easy consequence of the additivity is the following corollary. We leave the details to the reader as an exercise.

[intsubcor] If \(f \in {\mathcal{R}}[a,b]\) and \([c,d] \subset [a,b]\), then the restriction \(f|_{[c,d]}\) is in \({\mathcal{R}}[c,d]\).

#### Linearity and monotonicity

Let \(f\) and \(g\) be in \({\mathcal{R}}[a,b]\) and \(\alpha \in {\mathbb{R}}\).

- \(\alpha f\) is in \({\mathcal{R}}[a,b]\) and \[\int_a^b \alpha f(x) ~dx = \alpha \int_a^b f(x) ~dx .\]
- \(f+g\) is in \({\mathcal{R}}[a,b]\) and \[\int_a^b \bigl( f(x)+g(x) \bigr) ~dx = \int_a^b f(x) ~dx + \int_a^b g(x) ~dx .\]

Let us prove the first item. First suppose \(\alpha \geq 0\). Let \(P\) be a partition of \([a,b]\). Let \(m_i := \inf \{ f(x) : x \in [x_{i-1},x_i] \}\) as usual. Since \(\alpha\) is nonnegative, we can move the multiplication by \(\alpha\) past the infimum, \[\inf \{ \alpha f(x) : x \in [x_{i-1},x_i] \} = \alpha \inf \{ f(x) : x \in [x_{i-1},x_i] \} = \alpha m_i .\] Therefore \[L(P,\alpha f) = \sum_{i=1}^n \alpha m_i \Delta_i = \alpha \sum_{i=1}^n m_i \Delta_i = \alpha L(P,f).\] Similarly \[U(P,\alpha f) = \alpha U(P,f) .\] Again, as \(\alpha \geq 0\) we may move multiplication by \(\alpha\) past the supremum. Hence, \[\begin{split} \underline{\int_a^b} \alpha f(x)~dx & = \sup \{ L(P,\alpha f) : \text{$P$ a partition of $[a,b]$} \} \\ & = \sup \{ \alpha L(P,f) : \text{$P$ a partition of $[a,b]$} \} \\ & = \alpha \sup \{ L(P,f) : \text{$P$ a partition of $[a,b]$} \} \\ & = \alpha \underline{\int_a^b} f(x)~dx . \end{split}\] Similarly we show \[\overline{\int_a^b} \alpha f(x)~dx = \alpha \overline{\int_a^b} f(x)~dx .\] The conclusion now follows for \(\alpha \geq 0\).

To finish the proof of the first item, we need to show that \(-f\) is Riemann integrable and \(\int_a^b - f(x)~dx = - \int_a^b f(x)~dx\). The proof of this fact is left as an exercise.

The proof of the second item in the proposition is also left as an exercise. It is not as trivial as it may appear at first glance.

We should note that the second item in the proposition does not hold with equality for the Darboux integrals. For arbitrary bounded functions \(f\) and \(g\) we only obtain \[%\overline{\int_a^b} \bigl(f(x)+g(x)\bigr)~dx \leq %\overline{\int_a^b}f(x)~dx+\overline{\int_a^b}g(x)~dx \overline{\int_a^b} (f+g) \leq \overline{\int_a^b}f+\overline{\int_a^b}g , \qquad \text{and} \qquad \underline{\int_a^b} (f+g) \geq \underline{\int_a^b}f+\underline{\int_a^b}g %\underline{\int_a^b} \bigl(f(x)+g(x)\bigr)~dx \geq %\underline{\int_a^b}f(x)~dx+\underline{\int_a^b}g(x)~dx .\] See .

Let \(f\) and \(g\) be in \({\mathcal{R}}[a,b]\) and let \(f(x) \leq g(x)\) for all \(x \in [a,b]\). Then \[\int_a^b f \leq \int_a^b g .\]

Let \(P = \{ x_0, x_1, \ldots, x_n \}\) be a partition of \([a,b]\). Then let \[m_i := \inf \{ f(x) : x \in [x_{i-1},x_i] \} \qquad \text{and} \qquad \widetilde{m}_i := \inf \{ g(x) : x \in [x_{i-1},x_i] \} .\] As \(f(x) \leq g(x)\), then \(m_i \leq \widetilde{m}_i\). Therefore, \[L(P,f) = \sum_{i=1}^n m_i \Delta x_i \leq \sum_{i=1}^n \widetilde{m}_i \Delta x_i = L(P,g) .\] We take the supremum over all \(P\) (see ) to obtain \[\underline{\int_a^b} f \leq \underline{\int_a^b} g .\] As \(f\) and \(g\) are Riemann integrable, the conclusion follows.

#### Continuous functions

Let us show that continuous functions are Riemann integrable. In fact we will show we can even allow some discontinuities. We start with a function continuous on the whole closed interval \([a,b]\).

[lemma:contint] If \(f \colon [a,b] \to {\mathbb{R}}\) is a continuous function, then \(f \in {\mathcal{R}}[a,b]\).

As \(f\) is continuous on a closed bounded interval, it is uniformly continuous. Let \(\epsilon > 0\) be given. Find a \(\delta > 0\) such that \(\left\lvert {x-y} \right\rvert < \delta\) implies \(\left\lvert {f(x)-f(y)} \right\rvert < \frac{\epsilon}{b-a}\).

Let \(P = \{ x_0, x_1, \ldots, x_n \}\) be a partition of \([a,b]\) such that \(\Delta x_i < \delta\) for all \(i = 1,2, \ldots, n\). For example, take \(n\) such that \(\frac{b-a}{n} < \delta\) and let \(x_i := \frac{i}{n}(b-a) + a\). Then for all \(x, y \in [x_{i-1},x_i]\) we have \(\left\lvert {x-y} \right\rvert \leq \Delta x_i < \delta\) and so \[f(x)-f(y) \leq \left\lvert {f(x)-f(y)} \right\rvert < \frac{\epsilon}{b-a} .\] As \(f\) is continuous on \([x_{i-1},x_i]\), it attains a maximum and a minimum on this interval. Let \(x\) be a point where \(f\) attains the maximum and \(y\) be a point where \(f\) attains the minimum. Then \(f(x) = M_i\) and \(f(y) = m_i\) in the notation from the definition of the integral. Therefore, \[M_i-m_i = f(x)-f(y) < \frac{\epsilon}{b-a} .\] And so \[\begin{split} \overline{\int_a^b} f - \underline{\int_a^b} f & \leq U(P,f) - L(P,f) \\ & = \left( \sum_{i=1}^n M_i \Delta x_i \right) - \left( \sum_{i=1}^n m_i \Delta x_i \right) \\ & = \sum_{i=1}^n (M_i-m_i) \Delta x_i \\ & < \frac{\epsilon}{b-a} \sum_{i=1}^n \Delta x_i \\ & = \frac{\epsilon}{b-a} (b-a) = \epsilon . \end{split}\] As \(\epsilon > 0\) was arbitrary, \[\overline{\int_a^b} f = \underline{\int_a^b} f ,\] and \(f\) is Riemann integrable on \([a,b]\).

The second lemma says that we need the function to only be “Riemann integrable inside the interval,” as long as it is bounded. It also tells us how to compute the integral.

[lemma:boundedimpriemann] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function that is Riemann integrable on \([a',b']\) for all \(a',b'\) such that \(a < a' < b' < b\). Then \(f \in {\mathcal{R}}[a,b]\). Furthermore, if \(a < a_n < b_n < b\) are such that \(\lim \, a_n = a\) and \(\lim \, b_n = b\), then \[\int_a^b f = \lim_{n \to \infty} \int_{a_n}^{b_n} f .\]

Let \(M > 0\) be a real number such that \(\left\lvert {f(x)} \right\rvert \leq M\). Pick two sequences of numbers \(a < a_n < b_n < b\) such that \(\lim\, a_n = a\) and \(\lim\, b_n = b\). Note \(M > 0\) and \((b-a) \geq (b_n-a_n)\). Thus \[-M(b-a) \leq -M(b_n-a_n) \leq \int_{a_n}^{b_n} f \leq M(b_n-a_n) \leq M(b-a) .\] Therefore the sequence of numbers \(\{ \int_{a_n}^{b_n} f \}_{n=1}^\infty\) is bounded and by has a convergent subsequence indexed by \(n_k\). Let us call \(L\) the limit of the subsequence \(\{ \int_{a_{n_k}}^{b_{n_k}} f \}_{k=1}^\infty\).

says that the lower and upper integral are additive and the hypothesis says that \(f\) is integrable on \([a_{n_k},b_{n_k}]\). Therefore \[\underline{\int_a^b} f = \underline{\int_a^{a_{n_k}}} f + \int_{a_{n_k}}^{b_{n_k}} f + \underline{\int_{b_{n_k}}^b} f \geq -M(a_{n_k}-a) + \int_{a_{n_k}}^{b_{n_k}} f - M(b-b_{n_k}) .\] We take the limit as \(k\) goes to \(\infty\) on the right-hand side, \[\underline{\int_a^b} f \geq -M\cdot 0 + L - M\cdot 0 = L .\]

Next we use additivity of the upper integral, \[\overline{\int_a^b} f = \overline{\int_a^{a_{n_k}}} f + \int_{a_{n_k}}^{b_{n_k}} f + \overline{\int_{b_{n_k}}^b} f \leq M(a_{n_k}-a) + \int_{a_{n_k}}^{b_{n_k}} f + M(b-b_{n_k}) .\] We take the same subsequence \(\{ \int_{a_{n_k}}^{b_{n_k}} f \}_{k=1}^\infty\) and take the limit to obtain \[\overline{\int_a^b} f \leq M\cdot 0 + L + M\cdot 0 = L .\] Thus \(\overline{\int_a^b} f = \underline{\int_a^b} f = L\) and hence \(f\) is Riemann integrable and \(\int_a^b f = L\). In particular, no matter what sequences \(\{ a_n \}\) and \(\{b_n\}\) we started with and what subsequence we chose, the \(L\) is the same number.

To prove the final statement of the lemma we use . We have shown that every convergent subsequence \(\{ \int_{a_{n_k}}^{b_{n_k}} f \}\) converges to \(L = \int_a^b f\). Therefore, the sequence \(\{ \int_{a_n}^{b_n} f \}\) is convergent and converges to \(L\).

We say a function \(f \colon [a,b] \to {\mathbb{R}}\) has *finitely many discontinuities* if there exists a finite set \(S := \{ x_1, x_2, \ldots, x_n \} \subset [a,b]\), and \(f\) is continuous at all points of \([a,b] \setminus S\).

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a bounded function with finitely many discontinuities. Then \(f \in {\mathcal{R}}[a,b]\).

We divide the interval into finitely many intervals \([a_i,b_i]\) so that \(f\) is continuous on the interior \((a_i,b_i)\). If \(f\) is continuous on \((a_i,b_i)\), then it is continuous and hence integrable on \([c_i,d_i]\) whenever \(a_i < c_i < d_i < b_i\). By the restriction of \(f\) to \([a_i,b_i]\) is integrable. By additivity of the integral (and ) \(f\) is integrable on the union of the intervals.

Sometimes it is convenient (or necessary) to change certain values of a function and then integrate. The next result says that if we change the values only at finitely many points, the integral does not change.

Let \(f \colon [a,b] \to {\mathbb{R}}\) be Riemann integrable. Let \(g \colon [a,b] \to {\mathbb{R}}\) be a function such that \(f(x) = g(x)\) for all \(x \in [a,b] \setminus S\), where \(S\) is a finite set. Then \(g\) is a Riemann integrable function and \[\int_a^b g = \int_a^b f.\]

Using additivity of the integral, we split up the interval \([a,b]\) into smaller intervals such that \(f(x) = g(x)\) holds for all \(x\) except at the endpoints (details are left to the reader).

Therefore, without loss of generality suppose \(f(x) = g(x)\) for all \(x \in (a,b)\). The proof follows by , and is left as an exercise.

## Exercises

Let \(f\) be in \({\mathcal{R}}[a,b]\). Prove that \(-f\) is in \({\mathcal{R}}[a,b]\) and \[\int_a^b - f(x) ~dx = - \int_a^b f(x) ~dx .\]

Let \(f\) and \(g\) be in \({\mathcal{R}}[a,b]\). Prove that \(f+g\) is in \({\mathcal{R}}[a,b]\) and \[\int_a^b \bigl( f(x)+g(x) \bigr) ~dx = \int_a^b f(x) ~dx + \int_a^b g(x) ~dx .\] Hint: Use to find a single partition \(P\) such that \(U(P,f)-L(P,f) < \nicefrac{\epsilon}{2}\) and \(U(P,g)-L(P,g) < \nicefrac{\epsilon}{2}\).

Let \(f \colon [a,b] \to {\mathbb{R}}\) be Riemann integrable. Let \(g \colon [a,b] \to {\mathbb{R}}\) be a function such that \(f(x) = g(x)\) for all \(x \in (a,b)\). Prove that \(g\) is Riemann integrable and that \[\int_a^b g = \int_a^b f.\]

Prove the *mean value theorem for integrals*. That is, prove that if \(f \colon [a,b] \to {\mathbb{R}}\) is continuous, then there exists a \(c \in [a,b]\) such that \(\int_a^b f = f(c)(b-a)\).

If \(f \colon [a,b] \to {\mathbb{R}}\) is a continuous function such that \(f(x) \geq 0\) for all \(x \in [a,b]\) and \(\int_a^b f = 0\). Prove that \(f(x) = 0\) for all \(x\).

If \(f \colon [a,b] \to {\mathbb{R}}\) is a continuous function for all \(x \in [a,b]\) and \(\int_a^b f = 0\). Prove that there exists a \(c \in [a,b]\) such that \(f(c) = 0\) (Compare with the previous exercise).

If \(f \colon [a,b] \to {\mathbb{R}}\) and \(g \colon [a,b] \to {\mathbb{R}}\) are continuous functions such that \(\int_a^b f = \int_a^b g\). Then show that there exists a \(c \in [a,b]\) such that \(f(c) = g(c)\).

Let \(f \in {\mathcal{R}}[a,b]\). Let \(\alpha, \beta, \gamma\) be arbitrary numbers in \([a,b]\) (not necessarily ordered in any way). Prove \[\int_\alpha^\gamma f = \int_\alpha^\beta f + \int_\beta^\gamma f .\] Recall what \(\int_a^b f\) means if \(b \leq a\).

Prove .

[exercise:easyabsint] Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is bounded and has finitely many discontinuities. Show that as a function of \(x\) the expression \(\left\lvert {f(x)} \right\rvert\) is bounded with finitely many discontinuities and is thus Riemann integrable. Then show \[\left\lvert {\int_a^b f(x)~dx} \right\rvert \leq \int_a^b \left\lvert {f(x)} \right\rvert~dx .\]

Show that the Thomae or popcorn function (see ) is Riemann integrable. Therefore, there exists a function discontinuous at all rational numbers (a dense set) that is Riemann integrable.

In particular, define \(f \colon [0,1] \to {\mathbb{R}}\) by \[f(x) := \begin{cases} \nicefrac{1}{k} & \text{ if $x=\nicefrac{m}{k}$ where $m,k \in {\mathbb{N}}$ and $m$ and $k$ have no common divisors,} \\ 0 & \text{ if $x$ is irrational}. \end{cases}\] Show \(\int_0^1 f = 0\).

If \(I \subset {\mathbb{R}}\) is a bounded interval, then the function \[\varphi_I(x) := \begin{cases} 1 & \text{if $x \in I$,} \\ 0 & \text{otherwise,} \end{cases}\] is called an *elementary step function*.

Let \(I\) be an arbitrary bounded interval (you should consider all types of intervals: closed, open, half-open) and \(a < b\), then using only the definition of the integral show that the elementary step function \(\varphi_I\) is integrable on \([a,b]\), and find the integral in terms of \(a\), \(b\), and the endpoints of \(I\).

When a function \(f\) can be written as \[f(x) = \sum_{k=1}^n \alpha_k \varphi_{I_k} (x)\] for some real numbers \(\alpha_1,\alpha_2, \ldots, \alpha_n\) and some bounded intervals \(I_1,I_2,\ldots,I_n\), then \(f\) is called a *step function*.

Using the previous exercise, show that a step function is integrable on any interval \([a,b]\). Furthermore, find the integral in terms of \(a\), \(b\), the endpoints of \(I_k\) and the \(\alpha_k\).

[exercise:boundedvariationintegrable] Let \(f \colon [a,b] \to {\mathbb{R}}\) be increasing. a) Show that \(f\) is Riemann integrable. Hint: Use a uniform partition; each subinterval of same length. b) Use part a to show that a decreasing function is Riemann integrable. c) Suppose \(h = f-g\) where \(f\) and \(g\) are increasing functions on \([a,b]\). Show that \(h\) is Riemann integrable ^{22}.

[exercise:hardabsint] Suppose \(f \in {\mathcal{R}}[a,b]\), then the function that takes \(x\) to \(\left\lvert {f(x)} \right\rvert\) is also Riemann integrable on \([a,b]\). Then show the same inequality as .

[exercise:upperlowerlinineq] Suppose \(f \colon [a,b] \to {\mathbb{R}}\) and \(g \colon [a,b] \to {\mathbb{R}}\) are bounded. a) Show \(\overline{\int_a^b} (f+g) \leq \overline{\int_a^b}f+\overline{\int_a^b}g\) and \(\underline{\int_a^b} (f+g) \geq \underline{\int_a^b}f+\underline{\int_a^b}g\). b) Find example \(f\) and \(g\) where the inequality is strict. Hint: \(f\) and \(g\) should not be Riemann integrable.

### Fundamental theorem of calculus

*Note: 1.5 lectures*

In this chapter we discuss and prove the *fundamental theorem of calculus*. The entirety of integral calculus is built upon this theorem, ergo the name. The theorem relates the seemingly unrelated concepts of integral and derivative. It tells us how to compute the antiderivative of a function using the integral and vice-versa.

#### First form of the theorem

Let \(F \colon [a,b] \to {\mathbb{R}}\) be a continuous function, differentiable on \((a,b)\). Let \(f \in {\mathcal{R}}[a,b]\) be such that \(f(x) = F'(x)\) for \(x \in (a,b)\). Then \[\int_a^b f = F(b)-F(a) .\]

It is not hard to generalize the theorem to allow a finite number of points in \([a,b]\) where \(F\) is not differentiable, as long as it is continuous. This generalization is left as an exercise.

Let \(P = \{ x_0, x_1, \ldots, x_n \}\) be a partition of \([a,b]\). For each interval \([x_{i-1},x_i]\), use the to find a \(c_i \in (x_{i-1},x_i)\) such that \[f(c_i) \Delta x_i = F'(c_i) (x_i - x_{i-1}) = F(x_i) - F(x_{i-1}) .\] Using the notation from the definition of the integral, we have \(m_i \leq f(c_i) \leq M_i\) and so \[m_i \Delta x_i \leq F(x_i) - F(x_{i-1}) \leq M_i \Delta x_i .\] We sum over \(i = 1,2, \ldots, n\) to get \[\sum_{i=1}^n m_i \Delta x_i \leq \sum_{i=1}^n \bigl(F(x_i) - F(x_{i-1}) \bigr) \leq \sum_{i=1}^n M_i \Delta x_i .\] In the middle sum, all the terms except the first and last cancel and we end up with \(F(x_n)-F(x_0) = F(b)-F(a)\). The sums on the left and on the right are the lower and the upper sum respectively. So \[L(P,f) \leq F(b)-F(a) \leq U(P,f) .\] We take the supremum of \(L(P,f)\) over all \(P\) and the left inequality yields \[\underline{\int_a^b} f \leq F(b)-F(a) .\] Similarly, taking the infimum of \(U(P,f)\) over all partitions \(P\) yields \[F(b)-F(a) \leq \overline{\int_a^b} f .\] As \(f\) is Riemann integrable, we have \[\int_a^b f = \underline{\int_a^b} f \leq F(b)-F(a) \leq \overline{\int_a^b} f = \int_a^b f .\] The inequalities must be equalities and we are done.

The theorem is used to compute integrals. Suppose we know that the function \(f(x)\) is a derivative of some other function \(F(x)\), then we can find an explicit expression for \(\int_a^b f\).

Suppose we are trying to compute \[\int_0^1 x^2 ~dx .\] We notice \(x^2\) is the derivative of \(\frac{x^3}{3}\). We use the fundamental theorem to write \[\int_0^1 x^2 ~dx = \frac{1^3}{3} - \frac{0^3}{3} = \frac{1}{3}.\]

#### Second form of the theorem

The second form of the fundamental theorem gives us a way to solve the differential equation \(F'(x) = f(x)\), where \(f\) is a known function and we are trying to find an \(F\) that satisfies the equation.

Let \(f \colon [a,b] \to {\mathbb{R}}\) be a Riemann integrable function. Define \[F(x) := \int_a^x f .\] First, \(F\) is continuous on \([a,b]\). Second, if \(f\) is continuous at \(c \in [a,b]\), then \(F\) is differentiable at \(c\) and \(F'(c) = f(c)\).

As \(f\) is bounded, there is an \(M > 0\) such that \(\left\lvert {f(x)} \right\rvert \leq M\) for all \(x \in [a,b]\). Suppose \(x,y \in [a,b]\) with \(x > y\). Then \[\left\lvert {F(x)-F(y)} \right\rvert = \left\lvert {\int_a^x f - \int_a^y f} \right\rvert = \left\lvert {\int_y^x f} \right\rvert \leq M\left\lvert {x-y} \right\rvert .\] By symmetry, the same also holds if \(x < y\). So \(F\) is Lipschitz continuous and hence continuous.

Now suppose \(f\) is continuous at \(c\). Let \(\epsilon > 0\) be given. Let \(\delta > 0\) be such that for \(x \in [a,b]\) \(\left\lvert {x-c} \right\rvert < \delta\) implies \(\left\lvert {f(x)-f(c)} \right\rvert < \epsilon\). In particular, for such \(x\) we have \[f(c)-\epsilon \leq f(x) \leq f(c) + \epsilon.\] Thus if \(x > c\), then \[\bigl(f(c)-\epsilon\bigr) (x-c) \leq \int_c^x f \leq \bigl(f(c) + \epsilon\bigr)(x-c).\] When \(c > x\), then the inequalities are reversed. Therefore, assuming \(c \not= x\) we get \[f(c)-\epsilon \leq \frac{\int_c^{x} f}{x-c} \leq f(c)+\epsilon .\] As \[\frac{F(x)-F(c)}{x-c} = \frac{\int_a^{x} f - \int_a^{c} f}{x-c} = \frac{\int_c^{x} f}{x-c} ,\] we have \[\left\lvert {\frac{F(x)-F(c)}{x-c} - f(c)} \right\rvert \leq \epsilon .\] The result follows. It is left to the reader to see why is it OK that we just have a non-strict inequality.

Of course, if \(f\) is continuous on \([a,b]\), then it is automatically Riemann integrable, \(F\) is differentiable on all of \([a,b]\) and \(F'(x) = f(x)\) for all \(x \in [a,b]\).

The second form of the fundamental theorem of calculus still holds if we let \(d \in [a,b]\) and define \[F(x) := \int_d^x f .\] That is, we can use any point of \([a,b]\) as our base point. The proof is left as an exercise.

Let us look at what a simple discontinuity can do. Take \(f(x) := -1\) if \(x < 0\), and \(f(x) := 1\) if \(x \geq 0\). Let \(F(x) := \int_0^x f\). It is not difficult to see that \(F(x) = \left\lvert {x} \right\rvert\). Notice that \(f\) is discontinuous at \(0\) and \(F\) is not differentiable at \(0\). However, the converse does not hold. Let us do another quick example. Let \(g(x) := 0\) if \(x \not= 0\), and \(g(0) = 1\). Letting \(G(x) := \int_0^x g\), we find that \(G(x) = 0\) for all \(x\). So \(g\) is discontinuous at \(0\), but \(G'(0)\) exists and is equal to 0.

A common misunderstanding of the integral for calculus students is to think of integrals whose solution cannot be given in closed-form as somehow deficient. This is not the case. Most integrals we write down are not computable in closed-form. Even some integrals that we consider in closed-form are not really such. For example, how does a computer find the value of \(\ln x\)? One way to do it is to note that we define the natural log as the antiderivative of \(\nicefrac{1}{x}\) such that \(\ln 1 = 0\). Therefore, \[\ln x := \int_1^x \nicefrac{1}{s}~ds .\] Then we can numerically approximate the integral. Morally, we did not really “simplify” \(\int_1^x \nicefrac{1}{s}~ds\) by writing down \(\ln x\). We simply gave the integral a name. If we require numerical answers, it is possible we end up doing the calculation by approximating an integral anyway.

Another common function defined by an integral that cannot be evaluated symbolically is the erf function, defined as \[\operatorname{erf}(x) := \frac{2}{\sqrt{\pi}} \int_0^x e^{-s^2} ~ds .\] This function comes up often in applied mathematics. It is simply the antiderivative of \(\left(\nicefrac{2}{\sqrt{\pi}}\right) e^{-x^2}\) that is zero at zero. The second form of the fundamental theorem tells us that we can write the function as an integral. If we wish to compute any particular value, we numerically approximate the integral.

#### Change of variables

A theorem often used in calculus to solve integrals is the change of variables theorem. Let us prove it now. Recall a function is continuously differentiable if it is differentiable and the derivative is continuous.

Let \(g \colon [a,b] \to {\mathbb{R}}\) be a continuously differentiable function. If \(g([a,b]) \subset [c,d]\) and \(f \colon [c,d] \to {\mathbb{R}}\) is continuous, then \[\int_a^b f\bigl(g(x)\bigr)\, g'(x)~ dx = \int_{g(a)}^{g(b)} f(s)~ ds .\]

As \(g\), \(g'\), and \(f\) are continuous, we know \(f\bigl(g(x)\bigr)\,g'(x)\) is a continuous function on \([a,b]\), therefore it is Riemann integrable.

Define \[F(y) := \int_{g(a)}^{y} f(s)~ds .\] By the second form of the fundamental theorem of calculus (using below) \(F\) is a differentiable function and \(F'(y) = f(y)\). We apply the chain rule and write \[\bigl( F \circ g \bigr)' (x) = F'\bigl(g(x)\bigr) g'(x) = f\bigl(g(x)\bigr) g'(x) .\] We note that \(F\bigl(g(a)\bigr) = 0\) and we use the first form of the fundamental theorem to obtain \[\int_{g(a)}^{g(b)} f(s)~ds = F\bigl(g(b)\bigr) = F\bigl(g(b)\bigr)-F\bigl(g(a)\bigr) = \int_a^b \bigl( F \circ g \bigr)' (x) ~dx = \int_a^b f\bigl(g(x)\bigr) g'(x) ~dx . %FIXME \qedhere\]

The change of variables theorem is often used to solve integrals by changing them to integrals that we know or that we can solve using the fundamental theorem of calculus.

From an exercise, we know that the derivative of \(\sin(x)\) is \(\cos(x)\). Therefore we solve \[\int_0^{\sqrt{\pi}} x \cos(x^2) ~ dx = \int_0^\pi \frac{\cos(s)}{2} ~ ds = \frac{1}{2} \int_0^\pi \cos(s) ~ ds = \frac{ \sin(\pi) - \sin(0) }{2} = 0 .\]

However, beware that we must satisfy the hypotheses of the theorem. The following example demonstrates why we should not just move symbols around mindlessly. We must be careful that those symbols really make sense.

Suppose we write down \[\int_{-1}^{1} \frac{\ln \left\lvert {x} \right\rvert}{x} ~dx .\] It may be tempting to take \(g(x) := \ln \left\lvert {x} \right\rvert\). Then take \(g'(x) = \frac{1}{x}\) and try to write \[\int_{g(-1)}^{g(1)} s ~ds = \int_{0}^{0} s ~ds = 0.\] This “solution” is incorrect, and it does not say that we can solve the given integral. First problem is that \(\frac{\ln \left\lvert {x} \right\rvert}{x}\) is not continuous on \([-1,1]\). Second, \(\frac{\ln \left\lvert {x} \right\rvert}{x}\) is not even Riemann integrable on \([-1,1]\) (it is unbounded). The integral we wrote down simply does not make sense. Finally, \(g\) is not continuous on \([-1,1]\) either.

## Exercises

Compute \(\displaystyle \frac{d}{dx} \biggl( \int_{-x}^x e^{s^2}~ds \biggr)\).

Compute \(\displaystyle \frac{d}{dx} \biggl( \int_{0}^{x^2} \sin(s^2)~ds \biggr)\).

Suppose \(F \colon [a,b] \to {\mathbb{R}}\) is continuous and differentiable on \([a,b] \setminus S\), where \(S\) is a finite set. Suppose there exists an \(f \in {\mathcal{R}}[a,b]\) such that \(f(x) = F'(x)\) for \(x \in [a,b] \setminus S\). Show that \(\int_a^b f = F(b)-F(a)\).

[secondftc:exercise] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Let \(c \in [a,b]\) be arbitrary. Define \[F(x) := \int_c^x f .\] Prove that \(F\) is differentiable and that \(F'(x) = f(x)\) for all \(x \in [a,b]\).

Prove *integration by parts*. That is, suppose \(F\) and \(G\) are continuously differentiable functions on \([a,b]\). Then prove \[\int_a^b F(x)G'(x)~dx = F(b)G(b)-F(a)G(a) - \int_a^b F'(x)G(x)~dx .\]

Suppose \(F\) and \(G\) are continuously^{23} differentiable functions defined on \([a,b]\) such that \(F'(x) = G'(x)\) for all \(x \in [a,b]\). Using the fundamental theorem of calculus, show that \(F\) and \(G\) differ by a constant. That is, show that there exists a \(C \in {\mathbb{R}}\) such that \(F(x)-G(x) = C\).

The next exercise shows how we can use the integral to “smooth out” a non-differentiable function.

[exercise:smoothingout] Let \(f \colon [a,b] \to {\mathbb{R}}\) be a continuous function. Let \(\epsilon > 0\) be a constant. For \(x \in [a+\epsilon,b-\epsilon]\), define \[g(x) := \frac{1}{2\epsilon} \int_{x-\epsilon}^{x+\epsilon} f .\] a) Show that \(g\) is differentiable and find the derivative.

b) Let \(f\) be differentiable and fix \(x \in (a,b)\) (let \(\epsilon\) be small enough). What happens to \(g'(x)\) as \(\epsilon\) gets smaller?

c) Find \(g\) for \(f(x) := \left\lvert {x} \right\rvert\), \(\epsilon = 1\) (you can assume \([a,b]\) is large enough).

Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is continuous and \(\int_a^x f = \int_x^b f\) for all \(x \in [a,b]\). Show that \(f(x) = 0\) for all \(x \in [a,b]\).

Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is continuous and \(\int_a^x f = 0\) for all rational \(x\) in \([a,b]\). Show that \(f(x) = 0\) for all \(x \in [a,b]\).

A function \(f\) is an *odd function* if \(f(x) = -f(-x)\), and \(f\) is an *even function* if \(f(x) = f(-x)\). Let \(a > 0\). Assume \(f\) is continuous. Prove: a) If \(f\) is odd, then \(\int_{-a}^a f = 0\). b) If \(f\) is even, then \(\int_{-a}^a f = 2 \int_0^a f\).

a) Show that \(f(x) := \sin(\nicefrac{1}{x})\) is integrable on any interval (you can define \(f(0)\) to be anything). b) Compute \(\int_{-1}^1 \sin(\nicefrac{1}{x})\,dx\). (Mind the discontinuity)

§6] a) Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is increasing, by , \(f\) is Riemann integrable. Suppose \(f\) has a discontinuity at \(c \in (a,b)\), show that \(F(x) := \int_a^x f\) is not differentiable at \(c\).

b) In , you have constructed an increasing function \(f \colon [0,1] \to {\mathbb{R}}\) that is discontinuous at every \(x \in [0,1] \cap {\mathbb{Q}}\). Use this \(f\) to construct a function \(F(x)\) that is continuous on \([0,1]\), but not differentiable at every \(x \in [0,1] \cap {\mathbb{Q}}\).

### The logarithm and the exponential

*Note: 1 lecture (optional, requires the optional sections , , )*

We now have all that is required to finally properly define the exponential and the logarithm that you know from calculus so well. First recall that we have a good idea of what \(x^n\) means as long as \(n\) is a positive integer. Simply, \[x^n := \underbrace{x \cdot x \cdot \cdots \cdot x}_{\text{$n$ times}} .\] It makes sense to define \(x^0 := 1\). For negative integers we define \(x^{-n} := \nicefrac{1}{x^n}\). If \(x > 0\), we mentioned before that \(x^{1/n}\) is defined as the unique positive \(n\)th root. Finally for any rational number \(\nicefrac{n}{m}\), we define \[x^{n/m} := {\bigl(x^{1/m}\bigr)}^n .\] However, what do we mean by \(\sqrt{2}^{\sqrt{2}}\)? Or \(x^y\) in general? In particular, what is \(e^x\) for all \(x\)? And how do we solve \(y=e^x\) for \(x\)? This section answers these questions and more.

#### The logarithm

It is convenient to start with the logarithm. Let us show that a unique function with the right properties exists, and only then will we call it *the* logarithm.

There exists a unique function \(L \colon (0,\infty) \to {\mathbb{R}}\) such that

- [it:log:i] \(L(1) = 0\).
- [it:log:ii] \(L\) is differentiable and \(L'(x) = \nicefrac{1}{x}\).
- [it:log:iii] \(L\) is strictly increasing, bijective, and \[\lim_{x\to 0} L(x) = -\infty , \qquad \text{and} \qquad \lim_{x\to \infty} L(x) = \infty .\]
- [it:log:iv] \(L(xy) = L(x)+L(y)\) for all \(x,y \in (0,\infty)\).
- [it:log:v] If \(q\) is a rational number and \(x > 0\), then \(L(x^q) = q L(x)\).

To prove existence, let us define a candidate and show it satisfies all the properties. Define \[L(x) := \int_1^x \frac{1}{t}~dt .\]

Obviously [it:log:i] holds. Property [it:log:ii] holds via the fundamental theorem of calculus.

To prove property [it:log:iv], we change variables \(u=yt\) to obtain \[L(x) = \int_1^{x} \frac{1}{t}~dt = \int_y^{xy} \frac{1}{u}~du = \int_1^{xy} \frac{1}{u}~du - \int_1^{y} \frac{1}{u}~du = L(xy)-L(y) .\]

Property [it:log:ii] together with the fact that \(L'(x) = \nicefrac{1}{x} > 0\) for \(x > 0\), implies that \(L\) is strictly increasing and hence one-to-one. Let us show \(L\) is onto. As \(\nicefrac{1}{t} \geq \nicefrac{1}{2}\) when \(t \in [1,2]\), \[L(2) = \int_1^2 \frac{1}{t} ~dt \geq \nicefrac{1}{2} .\] By induction, [it:log:iv] implies that for \(n \in {\mathbb{N}}\) \[L(2^n) = L(2) + L(2) + \cdots + L(2) = n L(2) .\] Given any \(y > 0\), by the of the real numbers (notice \(L(2) > 0\)), there is an \(n \in {\mathbb{N}}\) such that \(L(2^n) > y\). By the there is an \(x_1 \in (1,2^n)\) such that \(L(x_1) = y\). We get \((0,\infty)\) is in the image of \(L\). As \(L\) is increasing, \(L(x) > y\) for all \(x > 2^n\), and so \[\lim_{x\to\infty} L(x) = \infty .\] Next \(0 = L(\nicefrac{x}{x}) = L(x) + L(\nicefrac{1}{x})\), and so \(L(x) = - L(\nicefrac{1}{x})\). Using \(x=2^{-n}\), we obtain as above that \(L\) achieves all negative numbers. And \[\lim_{x \to 0} L(x) = \lim_{x \to 0} -L(\nicefrac{1}{x}) = \lim_{x \to \infty} -L(x) = - \infty .\] In the limits, note that only \(x > 0\) are in the domain of \(L\).

Let us now prove [it:log:v]. As above, [it:log:iv] implies for \(n \in {\mathbb{N}}\) we have \(L(x^n) = n L(x)\). We already saw that \(L(x) = - L(\nicefrac{1}{x})\) so \(L(x^{-n}) = - L(x^n) = -n L(x)\). Then for \(m \in {\mathbb{N}}\) \[L(x) = L\Bigl({(x^{1/m})}^m\Bigr) = m L(x^{1/m}) .\] Putting everything together for \(n \in {\mathbb{Z}}\) and \(m \in {\mathbb{N}}\) we have \(L(x^{n/m}) = n L(x^{1/m}) = (\nicefrac{n}{m}) L(x)\).

Finally for uniqueness, let us use properties [it:log:i] and [it:log:ii]. Via the fundamental theorem of calculus \[L(x) = \int_1^x \frac{1}{t}~dt\] is the unique function such that \(L(1) = 0\) and \(L'(x) = \nicefrac{1}{x}\).

Having proved that there is a unique function with these properties we simply define the *logarithm* or sometimes called the *natural logarithm*: \[\ln(x) := L(x) .\] Often mathematicians write \(\log(x)\) instead of \(\ln(x)\), which is more familiar to calculus students.

#### The exponential

Just as with the logarithm we define the exponential via a list of properties.

There exists a unique function \(E \colon {\mathbb{R}}\to (0,\infty)\) such that

- [it:exp:i] \(E(0) = 1\).
- [it:exp:ii] \(E\) is differentiable and \(E'(x) = E(x)\).
- [it:exp:iii] \(E\) is strictly increasing, bijective, and \[\lim_{x\to -\infty} E(x) = 0 , \qquad \text{and} \qquad \lim_{x\to \infty} E(x) = \infty .\]
- [it:exp:iv] \(E(x+y) = E(x)E(y)\) for all \(x,y \in {\mathbb{R}}\).
- [it:exp:v] If \(q \in {\mathbb{Q}}\), then \(E(qx) = {E(x)}^q\).

Again, let us prove existence of such a function by defining a candidate, and prove that it satisfies all the properties. The \(L\) defined above is invertible. Let \(E\) be the inverse function of \(L\). Property [it:exp:i] is immediate.

Property [it:exp:ii] follows via the inverse function theorem, in particular : \(L\) satisfies all the hypotheses of the lemma, and hence \[E'(x) = \frac{1}{L'\bigl(E(x)\bigr)} = E(x) .\]

Let us look at property [it:exp:iii]. The function \(E\) is strictly increasing since \(E(x) > 0\) and \(E'(x) = E(x) > 0\). As \(E\) is the inverse of \(L\), it must also be bijective. To find the limits, we use that \(E\) is strictly increasing and onto \((0,\infty)\). For every \(M > 0\), there is an \(x_0\) such that \(E(x_0) = M\) and \(E(x) \geq M\) for all \(x \geq x_0\). Similarly for every \(\epsilon > 0\), there is an \(x_0\) such that \(E(x_0) = \epsilon\) and \(E(x) < \epsilon\) for all \(x < x_0\). Therefore, \[\lim_{n\to -\infty} E(x) = 0 , \qquad \text{and} \qquad \lim_{n\to \infty} E(x) = \infty .\]

To prove property [it:exp:iv] we use the corresponding property for the logarithm. Take \(x, y \in {\mathbb{R}}\). As \(L\) is bijective, find \(a\) and \(b\) such that \(x = L(a)\) and \(y = L(b)\). Then \[E(x+y) = E\bigl(L(a)+L(b)\bigr) = E\bigl(L(ab)\bigr) = ab = E(x)E(y) .\]

Property [it:exp:v] also follows from the corresponding property of \(L\). Given \(x \in {\mathbb{R}}\), let \(a\) be such that \(x = L(a)\) and \[E(qx) = E\bigl(qL(a)\bigr) E\bigl(L(a^q)\bigr) = a^q = {E(x)}^q .\]

Finally, uniqueness follows from [it:exp:i] and [it:exp:ii]. Let \(E\) and \(F\) be two functions satisfying [it:exp:i] and [it:exp:ii]. \[\frac{d}{dx} \Bigl( F(x)E(-x) \Bigr) = F'(x)E(-x) - E'(-x)F(x) = F(x)E(-x) - E(-x)F(x) = 0 .\] Therefore by , \(F(x)E(-x) = F(0)E(-0) = 1\) for all \(x \in {\mathbb{R}}\). Doing the computation with \(F = E\), we obtain \(E(x)E(-x) = 1\). Then \[0 = 1-1 = F(x)E(-x) - E(x)E(-x) = \bigl(F(x)-E(x)\bigr) E(-x) .\] Since \(E(x)E(-x) = 1\), then \(E(-x) \not= 0\) for all \(x\). So \(F(x)-E(x) = 0\) for all \(x\), and we are done.

Having proved \(E\) is unique, we define the *exponential* function as \[\exp(x) := E(x) .\]

We can now make sense of exponentiation \(x^y\) for arbitrary numbers when \(x > 0\). First suppose \(y \in {\mathbb{Q}}\). Then \[x^y = \exp\bigl(\ln(x^y)\bigr) = \exp\bigl(y\ln(x)\bigr) .\] Therefore when \(x > 0\) and \(y\) is irrational let us define \[x^y := \exp\bigl(y\ln(x)\bigr) .\] As \(\exp\) is continuous then \(x^y\) is a continuous function of \(y\). Therefore, we would obtain the same result had we taken a sequence of rational numbers \(\{ y_n \}\) approaching \(y\) and defined \(x^y = \lim\, x^{y_n}\).

Define the number \(e\) as \[e := \exp(1) .\] The number \(e\) is sometimes called *Euler’s number* or the *base of the natural logarithm*. We notice \[e^x = \exp\bigl(x \ln(e) \bigr) = \exp(x) .\] We have justified the notation \(e^x\) for \(\exp(x)\).

Finally, let us extend properties of logarithm and exponential to irrational powers. The proof is immediate.

Let \(x, y \in {\mathbb{R}}\).

- \(\exp(xy) = {\bigl(\exp(x)\bigr)}^y\).
- If \(x > 0\) then \(\ln(x^y) = y \ln (x)\).

## Exercises

Let \(y\) be any real number and \(b > 0\). Define \(f \colon (0,\infty) \to {\mathbb{R}}\) and \(g \colon {\mathbb{R}}\to {\mathbb{R}}\) as, \(f(x) := x^y\) and \(g(x) := b^x\). Show that \(f\) and \(g\) are differentiable and find their derivative.

Let \(b > 0\) be given.

a) Show that for every \(y > 0\), there exists a unique number \(x\) such that \(y = b^x\). Define the *logarithm base \(b\)*, \(\log_b \colon (0,\infty) \to {\mathbb{R}}\), by \(\log_b(y) := x\).

b) Show that \(\log_b(x) = \frac{\ln(x)}{\ln(b)}\).

c) Prove that if \(c > 0\), then \(\log_b(x) = \frac{\log_c(x)}{\log_c(b)}\).

d) Prove \(\log_b(xy) = \log_b(x)+\log_b(y)\), and \(\log_b(x^y) = y \log_b(x)\).

§3] Use to study the remainder term and show that for all \(x \in {\mathbb{R}}\) \[e^x = \sum_{n=0}^\infty \frac{x^n}{n!} .\] Hint: Do not differentiate the series term by term (unless you would prove that it works).

Use the geometric sum formula to show (for \(t\not= -1\)) \[1-t+t^2-\cdots+{(-1)}^n t^n = \frac{1}{1+t} - \frac

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[32]/p[4]/span[1], line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[32]/p[4]/span[2], line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Show \[e^x = \lim_{n\to\infty} {\left( 1 + \frac{x}{n} \right)}^n .\] Hint: Take the logarithm.

Note: The expression \({\left( 1 + \frac{x}{n} \right)}^n\) arises in compound interest calculations. It is the amount of money in a bank account after 1 year if 1 dollar was deposited initially at interest \(x\) and the interest was compounded \(n\) times during the year. Therefore \(e^x\) is the result of continuous compounding.

a) Prove that for \(n \in {\mathbb{N}}\) we have \[\sum_{k=2}^{n} \frac{1}{k} \leq \ln (n) \leq \sum_{k=1}^{n-1} \frac{1}{k} .\]

b) Prove that the limit \[\gamma := \lim_{n\to\infty} \left( \sum_{k=1}^{n} \frac{1}{k} - \ln (n) \right)\] exists. This constant is known as the *Euler-Mascheroni constant* ^{24}. It is not known if this constant is rational or not, it is approximately \(\gamma \approx 0.5772\).

Show \[\lim_{x\to\infty} \frac{\ln(x)}{x} = 0 .\]

Show that \(e^x\) is *convex*, in other words, show that if \(a \leq x \leq b\) then \(e^x \leq e^a \frac{b-x}{b-a} + e^b \frac{x-a}{b-a}\).

Using the logarithm find \[%\lim_{n\to\infty} {\left( 1 + \nicefrac{1}{n} \right)}^n = e . \lim_{n\to\infty} n^{1/n} .\]

Show that \(E(x) = e^x\) is the unique continuous function such that \(E(x+y) = E(x)E(y)\) and \(E(1) = e\). Similarly prove that \(L(x) = \ln(x)\) is the unique continuous function defined on positive \(x\) such that \(L(xy) = L(x)+L(y)\) and \(L(e) = 1\).

### Improper integrals

*Note: 2–3 lectures (optional section, can safely be skipped, requires the optional )*

Often it is necessary to integrate over the entire real line, or a infinite interval of the form \([a,\infty)\) or \((\infty,b]\). Also, we may wish to integrate functions defined on a finite interval \((a,b)\) but not bounded. Such functions are not Riemann integrable, but we may want to write down the integral anyway in the spirit of . These integrals are called *improper integrals*, and are limits of integrals rather than integrals themselves.

Suppose \(f \colon [a,b) \to {\mathbb{R}}\) is a function (not necessarily bounded) that is Riemann integrable on \([a,c]\) for all \(c < b\). We define \[\int_a^b f := \lim_{c \to b^-} \int_a^{c} f ,\] if the limit exists.

Suppose \(f \colon [a,\infty) \to {\mathbb{R}}\) is a function such that \(f\) is Riemann integrable on \([a,c]\) for all \(c < \infty\). We define \[\int_a^\infty f := \lim_{c \to \infty} \int_a^c f ,\] if the limit exists.

If the limit exists, we say the improper integral *converges*. If the limit does not exist, we say the improper integral *diverges*.

We similarly define improper integrals for the left hand endpoint, we leave this to the reader.

For a finite endpoint \(b\), using we see that if \(f\) is bounded, then we have defined nothing new. What is new is that we can apply this definition to unbounded functions. The following set of examples is so useful that we state it as a proposition.

[impropriemann:ptest] The improper integral \[\int_1^\infty \frac{1}{x^p} ~dx\] converges to \(\frac{1}{p-1}\) if \(p > 1\) and diverges if \(0 < p \leq 1\).

The improper integral \[\int_0^1 \frac{1}{x^p} ~dx\] converges to \(\frac{1}{1-p}\) if \(0 < p < 1\) and diverges if \(p \geq 1\).

The proof follows by application of the fundamental theorem of calculus. Let us do the proof for \(p > 1\) for the infinite right endpoint, and we leave the rest to the reader. Hint: You should handle \(p=1\) separately.

Suppose \(p > 1\). Then \[\int_1^b \frac{1}{x^p} ~dx = \int_1^b x^{-p} ~dx = \frac{b^{-p+1}}{-p+1} - \frac{1^{-p+1}}{-p+1} = - \frac{1}{(p-1)b^{p-1}} + \frac{1}{p-1} .\] As \(p > 1\), then \(p-1 > 0\). Taking the limit as \(b \to \infty\) we obtain that \(\frac{1}{b^{p-1}}\) goes to 0, and the result follows.

We state the following proposition for just one type of improper integral, though the proof is straight forward and the same for other types of improper integrals.

[impropriemann:tail] Let \(f \colon [a,\infty) \to {\mathbb{R}}\) be a function that is Riemann integrable on \([a,b]\) for all \(b > a\). Given any \(b > a\), \(\int_b^\infty f\) converges if and only if \(\int_a^\infty f\) converges, in which case \[\int_a^\infty f = \int_a^b f + \int_b^\infty f .\]

Let \(c > b\). Then \[\int_a^c f = \int_a^b f + \int_b^c f .\] Taking the limit \(c \to \infty\) finishes the proof.

Nonnegative functions are easier to work with as the following proposition demonstrates. The exercises will show that this proposition holds only for nonnegative functions. Analogues of this proposition exist for all the other types of improper limits are left to the student.

[impropriemann:possimp] Suppose \(f \colon [a,\infty) \to {\mathbb{R}}\) is nonnegative (\(f(x) \geq 0\) for all \(x\)) and such that \(f\) is Riemann integrable on \([a,b]\) for all \(b > a\).

- \[\int_a^\infty f = \sup \left\{ \int_a^x f : x \geq a \right\} .\]
- Suppose \(\{ x_n \}\) is a sequence with \(\lim\, x_n = \infty\). Then \(\int_a^\infty f\) converges if and only if \(\lim\, \int_a^{x_n} f\) exists, in which case \[\int_a^\infty f = \lim_{n\to\infty} \int_a^{x_n} f .\]

In the first item we allow for the value of \(\infty\) in the supremum indicating that the integral diverges to infinity.

Let us start with the first item. Notice that as \(f\) is nonnegative, then \(\int_a^x f\) is increasing as a function of \(x\). If the supremum is infinite, then for every \(M \in {\mathbb{R}}\) we find \(N\) such that \(\int_a^N f \geq M\). As \(\int_a^x f\) is increasing then \(\int_a^x f \geq M\) for all \(x \geq N\). So \(\int_a^\infty f\) diverges to infinity.

Next suppose the supremum is finite, say \(A = \sup \left\{ \int_a^x f : x \geq a \right\}\). For every \(\epsilon > 0\), we find an \(N\) such that \(A - \int_a^N f < \epsilon\). As \(\int_a^x f\) is increasing, then \(A - \int_a^x f < \epsilon\) for all \(x \geq N\) and hence \(\int_a^\infty f\) converges to \(A\).

Let us look at the second item. If \(\int_a^\infty f\) converges then every sequence \(\{ x_n \}\) going to infinity works. The trick is proving the other direction. Suppose \(\{ x_n \}\) is such that \(\lim\, x_n = \infty\) and \[\lim_{n\to\infty} \int_a^{x_n} f = A\] converges. Given \(\epsilon > 0\), pick \(N\) such that for all \(n \geq N\) we have \(A - \epsilon < \int_a^{x_n} f < A + \epsilon\). Because \(\int_a^x f\) is increasing as a function of \(x\), we have that for all \(x \geq x_N\) \[A - \epsilon < \int_a^{x_N} \leq \int_a^x f .\] As \(\{ x_n \}\) goes to \(\infty\), then for any given \(x\), there is an \(x_m\) such that \(m \geq N\) and \(x \leq x_m\). Then \[\int_a^{x} f \leq \int_a^{x_m} f < A + \epsilon .\] In particular, for all \(x \geq x_N\) we have \(\left\lvert {\int_a^{x} f - A} \right\rvert < \epsilon\).

Let \(f \colon [a,\infty) \to {\mathbb{R}}\) and \(g \colon [a,\infty) \to {\mathbb{R}}\) be functions that are Riemann integrable on \([a,b]\) for all \(b > a\). Suppose that for all \(x \geq a\) we have \[\left\lvert {f(x)} \right\rvert \leq g(x) .\]

- If \(\int_a^\infty g\) converges, then \(\int_a^\infty f\) converges, and in this case \(\left\lvert {\int_a^\infty f} \right\rvert \leq \int_a^\infty g\).
- If \(\int_a^\infty f\) diverges, then \(\int_a^\infty g\) diverges.

Let us start with the first item. For any \(b\) and \(c\), such that \(a \leq b \leq c\), we have \(-g(x) \leq f(x) \leq g(x)\), and so \[\int_b^c -g \leq \int_b^c f \leq \int_b^c g .\] In other words, \(\left\lvert {\int_b^c f} \right\rvert \leq \int_b^c g\).

Let \(\epsilon > 0\) be given. Because of we have \[\int_a^\infty g = \int_a^b g + \int_b^\infty g .\] As \(\int_a^b g\) goes to \(\int_a^\infty g\) as \(b\) goes to infinity, then \(\int_b^\infty g\) goes to 0 as \(b\) goes to infinity. Choose \(B\) such that \[\int_B^\infty g < \epsilon .\] As \(g\) is nonnegative, then if \(B \leq b < c\), then \(\int_b^c g < \epsilon\) as well. Let \(\{ x_n \}\) be a sequence going to infinity. Let \(M\) be such that \(x_n \geq B\) for all \(n \geq M\). Take \(n, m \geq M\), with \(x_n \leq x_m\), \[\left\lvert {\int_a^{x_m} f - \int_a^{x_n} f} \right\rvert = \left\lvert {\int_{x_n}^{x_m} f} \right\rvert \leq \int_{x_n}^{x_m} g < \epsilon .\] Therefore the sequence \(\{ \int_a^{x_n} f \}_{n=1}^\infty\) is Cauchy and hence converges.

We need to show that the limit is unique. Suppose \(\{ x_n \}\) is a sequence converging to infinity such that \(\{ \int_a^{x_n} f \}\) converges to \(L_1\), and \(\{ y_n \}\) is a sequence converging to infinity is such that \(\{ \int_a^{y_n} f \}\) converges to \(L_2\). Then there must be some \(n\) such that \(\left\lvert {\int_a^{x_n} f - L_1} \right\rvert < \epsilon\) and \(\left\lvert {\int_a^{y_n} f - L_2} \right\rvert < \epsilon\). We can also suppose \(x_n \geq B\) and \(y_n \geq B\). Then \[\left\lvert {L_1 - L_2} \right\rvert \leq \left\lvert {L_1 - \int_a^{x_n} f} \right\rvert + \left\lvert {\int_a^{x_n} f- \int_a^{y_n} f} \right\rvert + \left\lvert {\int_a^{y_n} f - L_2} \right\rvert < \epsilon + \left\lvert {\int_{x_n}^{y_n} f} \right\rvert + \epsilon < 3 \epsilon.\] As \(\epsilon > 0\) was arbitrary, \(L_1 = L_2\), and hence \(\int_a^\infty f\) converges. Above we have shown that \(\left\lvert {\int_a^c f} \right\rvert \leq \int_a^c g\) for all \(c > a\). By taking the limit \(c \to \infty\), the first item is proved.

The second item is simply a contrapositive of the first item.

The improper integral \[\int_0^\infty \frac{\sin(x^2)(x+2)}{x^3+1} ~dx\] converges.

Proof: First observe we simply need to show that the integral converges when going from 1 to infinity. For \(x \geq 1\) we obtain \[\left\lvert {\frac{\sin(x^2)(x+2)}{x^3+1}} \right\rvert \leq \frac{x+2}{x^3+1} \leq \frac{x+2}{x^3} \leq \frac{x+2x}{x^3} \leq \frac{3}{x^2} .\] Then \[3 \int_1^\infty \frac{1}{x^2}~dx = \lim_{c\to\infty} \int_1^c \frac{3}{x^2} ~dx.\] So the integral converges.

You should be careful when doing formal manipulations with improper integrals. For example, \[\int_2^\infty \frac{2}{x^2-1}~dx\] converges via the comparison test again using \(\frac{1}{x^2}\). However, if you succumb to the temptation to write \[\frac{2}{x^2-1} = \frac{1}{x-1} - \frac{1}{x+1}\] and try to integrate each part separately, you will not succeed. It is *not* true that you can split the improper integral in two; you cannot split the limit. \[\begin{split} \int_2^\infty \frac{2}{x^2-1} ~dx &= \lim_{b\to \infty} \int_2^b \frac{2}{x^2-1} ~dx \\ &= \lim_{b\to \infty} \left( \int_2^b \frac{1}{x-1}~dx - \int_2^b \frac{1}{x+1}~dx \right) \\ &\not= \int_2^\infty \frac{1}{x-1}~dx - \int_2^\infty \frac{1}{x+1}~dx . \end{split}\] The last line in the computation does not even make sense. Both of the integrals there diverge to infinity since we can apply the comparison test appropriately with \(\nicefrac{1}{x}\). We get \(\infty - \infty\).

Now let us suppose that we need to take limits at both endpoints.

Suppose \(f \colon (a,b) \to {\mathbb{R}}\) is a function that is Riemann integrable on \([c,d]\) for all \(c\), \(d\) such that \(a < c < d < b\), then we define \[\int_a^b f := \lim_{c \to a^+} \, \lim_{d \to b^-} \, \int_{c}^{d} f ,\] if the limits exist.

Suppose \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a function such that \(f\) is Riemann integrable on all finite intervals \([a,b]\). Then we define \[\int_{-\infty}^\infty f := \lim_{c \to -\infty} \, \lim_{d \to \infty} \, \int_c^d f ,\] if the limits exist.

We similarly define improper integrals with one infinite and one finite improper endpoint, we leave this to the reader.

One ought to always be careful about double limits. The definition given above says that we first take the limit as \(d\) goes to \(b\) or \(\infty\) for a fixed \(c\), and then we take the limit in \(c\). We will have to prove that in this case it does not matter which limit we compute first.

Let us see an example: \[\int_{-\infty}^\infty \frac{1}{1+x^2} ~ dx = \lim_{a \to -\infty} \, \lim_{b \to \infty} \, \int_{a}^b \frac{1}{1+x^2} ~ dx = \lim_{a \to -\infty} \, \lim_{b \to \infty} \bigl( \arctan(b) - \arctan(a) \bigr) = \pi .\]

In the definition the order of the limits can always be switched if they exist. Let us prove this fact only for the infinite limits.

If \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is a function integrable on every interval. Then \[\lim_{a \to -\infty} \, \lim_{b \to \infty} \, \int_a^b f \quad \text{converges if and only if} \qquad \lim_{b \to \infty} \, \lim_{a \to -\infty} \, \int_a^b f \quad \text{converges,}\] in which case the two expressions are equal. If either of the expressions converges then the improper integral converges and \[\lim_{a\to\infty} \int_{-a}^a f = \int_{-\infty}^\infty f .\]

Without loss of generality assume \(a < 0\) and \(b > 0\). Suppose the first expression converges. Then \[\begin{split} \lim_{a \to -\infty} \, \lim_{b \to \infty} \, \int_a^b f & = \lim_{a \to -\infty} \, \lim_{b \to \infty} \left( \int_a^0 f + \int_0^b f \right) = \left( \lim_{a \to -\infty} \int_a^0 f \right) + \left( \lim_{b \to \infty} \int_0^b f \right) \\ & = \lim_{b \to \infty} \left( \left( \lim_{a \to -\infty} \int_a^0 f \right) + \int_0^b f \right) = \lim_{b \to \infty} \, \lim_{a \to -\infty} \left( \int_a^0 f + \int_0^b f \right) . \end{split}\] Similar computation shows the other direction. Therefore, if either expression converges then the improper integral converges and \[\begin{split} \int_{-\infty}^\infty f = \lim_{a \to -\infty} \, \lim_{b \to \infty} \, \int_a^b f & = \left( \lim_{a \to -\infty} \int_a^0 f \right) + \left( \lim_{b \to \infty} \int_0^b f \right) \\ & = \left( \lim_{a \to \infty} \int_{-a}^0 f \right) + \left( \lim_{a \to \infty} \int_0^a f \right) = \lim_{a \to \infty} \left( \int_{-a}^0 f + \int_0^a f \right) = \lim_{a \to \infty} \int_{-a}^a f . \end{split}\]

On the other hand, you must be careful to take the limits independently before you know convergence. Let \(f(x) = \frac{x}{\left\lvert {x} \right\rvert}\) for \(x \not= 0\) and \(f(0) = 0\). If \(a < 0\) and \(b > 0\), then \[\int_{a}^b f = \int_{a}^0 f + \int_{0}^b f = a+b .\] For any fixed \(a < 0\) the limit as \(b \to \infty\) is infinite, so even the first limit does not exist, and hence the improper integral \(\int_{-\infty}^\infty f\) does not converge. On the other hand if \(a > 0\), then \[\int_{-a}^{a} f = (-a)+a = 0 .\] Therefore, \[\lim_{a\to\infty} \int_{-a}^{a} f = 0 .\]

An example to keep in mind for improper integrals is the so-called *sinc function* ^{25}. This function comes up quite often in both pure and applied mathematics. Define \[\operatorname{sinc}(x) = \begin{cases} \frac{\sin(x)}{x} & \text{if $x \not= 0$} , \\ 0 & \text{if $x = 0$} . \end{cases}\]

It is not difficult to show that the sinc function is continuous at zero, but that is not important right now. What is important is that \[\int_{-\infty}^\infty \operatorname{sinc}(x) ~dx = \pi , \qquad \text{while} \qquad \int_{-\infty}^\infty \left\lvert {\operatorname{sinc}(x)} \right\rvert ~dx = \infty .\] The integral of the sinc function is a continuous analogue of the alternating harmonic series \(\sum \nicefrac{{(-1)}^n}{n}\), while the absolute value is like the regular harmonic series \(\sum \nicefrac{1}{n}\). In particular, the fact that the integral converges must be done directly rather than using comparison test.

We will not prove the first statement exactly. Let us simply prove that the integral of the sinc function converges, but we will not worry about the exact limit. Because \(\frac{\sin(-x)}{-x} = \frac{\sin(x)}{x}\), it is enough to show that \[\int_{2\pi}^\infty \frac{\sin(x)}{x}~dx\] converges. We also avoid \(x=0\) this way to make our life simpler.

For any \(n \in {\mathbb{N}}\), we have that for \(x \in [\pi 2n, \pi (2n+1)]\) \[\frac{\sin(x)}{\pi (2n+1)} \leq \frac{\sin(x)}{x} \leq \frac{\sin(x)}{\pi 2n} ,\] as \(\sin(x) \geq 0\). On \(x \in [\pi (2n+1), \pi (2n+2)]\) \[\frac{\sin(x)}{\pi (2n+1)} \leq \frac{\sin(x)}{x} \leq \frac{\sin(x)}{\pi (2n+2)} ,\] as \(\sin(x) \leq 0\).

Via the fundamental theorem of calculus, \[\frac{2}{\pi (2n+1)} = \int_{\pi 2n}^{\pi (2n+1)} \frac{\sin(x)}{\pi (2n+1)} ~dx \leq \int_{\pi 2n}^{\pi (2n+1)} \frac{\sin(x)}{x} ~dx \leq \int_{\pi 2n}^{\pi (2n+1)} \frac{\sin(x)}{\pi 2n} ~dx = \frac{1}{\pi n} .\] Similarly \[\frac{-2}{\pi (2n+1)} \leq \int_{\pi (2n+1)}^{\pi (2n+2)} \frac{\sin(x)}{x} ~dx \leq \frac{-1}{\pi (n+1)} .\] Putting the two together we have \[0 = \frac{2}{\pi (2n+1)} - \frac{2}{\pi (2n+1)} + \leq \int_{2\pi n}^{2\pi (n+1)} \frac{\sin(x)}{x} ~dx \leq \frac{1}{\pi n} - \frac{1}{\pi (n+1)} = \frac{1}{\pi n(n+1)} .\] Let \(M > 2\pi\) be arbitrary, and let \(k \in {\mathbb{N}}\) be the largest integer such that \(2k\pi \leq M\). Then \[\int_{2\pi}^M \frac{\sin(x)}{x}~dx = \int_{2\pi}^{2k\pi} \frac{\sin(x)}{x} ~dx + \int_{2k\pi}^{M} \frac{\sin(x)}{x} ~dx .\] For \(x \in [2k\pi,M]\) we have \(\frac{-1}{2k\pi} \leq \frac{\sin(x)}{x} \leq \frac{1}{2k\pi}\), and so \[\left\lvert {\int_{2k\pi}^{M} \frac{\sin(x)}{x} ~dx } \right\rvert \leq \frac{M-2k\pi}{2k\pi} \leq \frac{1}{k} .\] As \(k\) is the largest \(k\) such that \(2k\pi \leq M\), this term goes to zero as \(M\) goes to infinity.

Next \[0 \leq \int_{2\pi}^{2k\pi} \frac{\sin(x)}{x} \leq \sum_{n=1}^{k-1} \frac{1}{\pi n(n+1)} ,\] and this series converges as \(k \to \infty\).

Putting the two statements together we obtain \[\int_{2\pi}^\infty \frac{\sin(x)}{x} ~dx \leq \sum_{n=1}^{\infty} \frac{1}{\pi n(n+1)} < \infty .\]

The double sided integral of sinc also exists as noted above. We leave the other statement—that the integral of the absolute value of the sinc function diverges—as an exercise.

#### Integral test for series

It can be very useful to apply the fundamental theorem of calculus in proving a series is summable and to estimate its sum.

Suppose \(f \colon [k,\infty) \to {\mathbb{R}}\) is a decreasing nonnegative function where \(k \in {\mathbb{Z}}\). Then \[\sum_{n=k}^\infty f(n) \quad \text{converges if and only if} \qquad \int_k^\infty f \quad \text{converges}.\] In this case \[\int_k^\infty f \leq \sum_{n=k}^\infty f(n) \leq f(k)+ \int_k^\infty f .\]

By , \(f\) is integrable on every interval \([k,b]\) for all \(b > k\), so the statement of the theorem makes sense without additional hypotheses of integrability.

Let \(\epsilon > 0\) be given. And suppose \(\int_k^\infty f\) converges. Let \(\ell, m \in {\mathbb{Z}}\) be such that \(m > \ell \geq k\). Because \(f\) is decreasing we have \(\int_{n}^{n+1} f \leq f(n) \leq \int_{n-1}^{n} f\). Therefore \[\label{impropriemann:eqseries} \int_\ell^m f = \sum_{n=\ell}^{m-1} \int_{n}^{n+1} f \leq \sum_{n=\ell}^{m-1} f(n) \leq f(\ell) + \sum_{n=\ell+1}^{m-1} \int_{n-1}^{n} f \leq f(\ell)+ \int_\ell^{m-1} f .\] As before, since \(f\) is positive then there exists an \(L \in {\mathbb{N}}\) such that if \(\ell \geq L\), then \(\int_\ell^{m} f < \nicefrac{\epsilon}{2}\) for all \(m \geq \ell\). We note \(f\) must decrease to zero (why?). So let us also suppose that for \(\ell \geq L\) we have \(f(\ell) < \nicefrac{\epsilon}{2}\). For such \(\ell\) and \(m\) we have via [impropriemann:eqseries] \[\sum_{n=\ell}^{m} f(n) \leq f(\ell)+ \int_\ell^{m} f < \nicefrac{\epsilon}{2} + \nicefrac{\epsilon}{2} = \epsilon .\] The series is therefore Cauchy and thus converges. The estimate in the proposition is obtained by letting \(m\) go to infinity in [impropriemann:eqseries] with \(\ell = k\).

Conversely suppose \(\int_k^\infty f\) diverges. As \(f\) is positive then by , the sequence \(\{ \int_k^m f \}_{m=k}^\infty\) diverges to infinity. Using [impropriemann:eqseries] with \(\ell = k\) we find \[\int_k^m f \leq \sum_{n=k}^{m-1} f(n) .\] As the left hand side goes to infinity as \(m \to \infty\), so does the right hand side.

Let us show \(\sum_{n=1}^\infty \frac{1}{n^2}\) exists and let us estimate its sum to within 0.01. As this series is the \(p\)-series for \(p=2\), we already know it converges, but we have only very roughly estimated its sum.

Using fundamental theorem of calculus we find that for \(k \in {\mathbb{N}}\) we have \[\int_{k}^\infty \frac{1}{x^2}~dx = \frac{1}{k} .\] In particular, the series must converge. But we also have that \[\frac{1}{k} = \int_k^\infty \frac{1}{x^2}~dx \leq \sum_{n=k}^\infty \frac{1}{n^2} \leq \frac{1}{k^2} + \int_k^\infty \frac{1}{x^2}~dx = \frac{1}{k^2} + \frac{1}{k} .\] Adding the partial sum up to \(k-1\) we get \[\frac{1}{k} + \sum_{n=1}^{k-1} \frac{1}{n^2} \leq \sum_{n=1}^\infty \frac{1}{n^2} \leq \frac{1}{k^2} + \frac{1}{k} + \sum_{n=1}^{k-1} \frac{1}{n^2} .\] In other words, \(\nicefrac{1}{k} + \sum_{n=1}^{k-1} \nicefrac{1}{n^2}\) is an estimate for the sum to within \(\nicefrac{1}{k^2}\). Therefore, if we wish to find the sum to within 0.01, we note \(\nicefrac{1}{{10}^2} = 0.01\). We obtain \[1.6397\ldots \approx \frac{1}{10} + \sum_{n=1}^{9} \frac{1}{n^2} \leq \sum_{n=1}^\infty \frac{1}{n^2} \leq \frac{1}{100} + \frac{1}{10} + \sum_{n=1}^{9} \frac{1}{n^2} \approx 1.6497\ldots .\] The actual sum is \(\nicefrac{\pi^2}{6} \approx 1.6449\ldots\).

## Exercises

Finish the proof of .

Find out for which \(a \in {\mathbb{R}}\) does \(\sum\limits_{n=1}^\infty e^{an}\) converge. When the series converges, find an upper bound for the sum.

a) Estimate \(\sum\limits_{n=1}^\infty \frac{1}{n(n+1)}\) correct to within 0.01 using the integral test. b) Compute the limit of the series exactly and compare. Hint: the sum telescopes.

Prove \[\int_{-\infty}^\infty \left\lvert {\operatorname{sinc}(x)} \right\rvert~dx = \infty .\] Hint: again, it is enough to show this on just one side.

Can you interpret \[\int_{-1}^1 \frac{1}{\sqrt{\left\lvert {x} \right\rvert}}~dx\] as an improper integral? If so, compute its value.

Take \(f \colon [0,\infty) \to {\mathbb{R}}\), Riemann integrable on every interval \([0,b]\), and such that there exist \(M\), \(a\), and \(T\), such that \(\left\lvert {f(t)} \right\rvert \leq M e^{at}\) for all \(t \geq T\). Show that the *Laplace transform* of \(f\) exists. That is, for every \(s > a\) the following integral converges: \[F(s) := \int_{0}^\infty f(t) e^{-st} ~dt .\]

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) be a Riemann integrable function on every interval \([a,b]\), and such that \(\int_{-\infty}^\infty \left\lvert {f(x)} \right\rvert~dx < \infty\). Show that the *Fourier sine and cosine transforms* exist. That is, for every \(\omega \geq 0\) the following integrals converge \[F^s(\omega) := \frac{1}{\pi} \int_{-\infty}^\infty f(t) \sin(\omega t) ~dt , \qquad F^c(\omega) := \frac{1}{\pi} \int_{-\infty}^\infty f(t) \cos(\omega t) ~dt .\] Furthermore, show that \(F^s\) and \(F^c\) are bounded functions.

Suppose \(f \colon [0,\infty) \to {\mathbb{R}}\) is Riemann integrable on every interval \([0,b]\). Show that \(\int_0^\infty f\) converges if and only if for every \(\epsilon > 0\) there exists an \(M\) such that if \(M \leq a < b\) then \(\left\lvert {\int_a^b f} \right\rvert < \epsilon\).

Suppose \(f \colon [0,\infty) \to {\mathbb{R}}\) is nonnegative and *decreasing*. a) Show that if \(\int_0^\infty f < \infty\), then \(\lim\limits_{x\to\infty} f(x) = 0\). b) Show that the converse does not hold.

Find an example of an *unbounded* continuous function \(f \colon [0,\infty) \to {\mathbb{R}}\) that is nonnegative and such that \(\int_0^\infty f < \infty\). Note that this means that \(\lim_{x\to\infty} f(x)\) does not exist; compare previous exercise. Hint: on each interval \([k,k+1]\), \(k \in {\mathbb{N}}\), define a function whose integral over this interval is less than say \(2^{-k}\).

Find an example of a function \(f \colon [0,\infty) \to {\mathbb{R}}\) integrable on all intervals such that \(\lim_{n\to\infty} \int_0^n f\) converges as a limit of a sequence, but such that \(\int_0^\infty f\) does not exist. Hint: for all \(n\in {\mathbb{N}}\), divide \([n,n+1]\) into two halves. In one half make the function negative, on the other make the function positive.

Show that if \(f \colon [1,\infty) \to {\mathbb{R}}\) is such that \(g(x) := x^2 f(x)\) is a bounded function, then \(\int_1^\infty f\) converges.

It is sometimes desirable to assign a value to integrals that normally cannot be interpreted as even improper integrals, e.g. \(\int_{-1}^1 \nicefrac{1}{x}~dx\). Suppose \(f \colon [a,b] \to {\mathbb{R}}\) is a function and \(a < c < b\), where \(f\) is Riemann integrable on all intervals \([a,c-\epsilon]\) and \([c+\epsilon,b]\) for all \(\epsilon > 0\). Define the *Cauchy principal value* of \(\int_a^b f\) as \[p.v.\!\int_a^b f := \lim_{\epsilon\to 0^+} \left( \int_a^{c-\epsilon} f + \int_{c+\epsilon}^b f \right) ,\] if the limit exists.

a) Compute \(p.v.\!\int_{-1}^1 \nicefrac{1}{x}~dx\).

b) Compute \(\lim_{\epsilon\to 0^+} ( \int_{-1}^{-\epsilon} \nicefrac{1}{x}~dx + \int_{2\epsilon}^1 \nicefrac{1}{x}~dx )\) and show it is not equal to the principal value.

c) Show that if \(f\) is integrable on \([a,b]\), then \(p.v.\!\int_a^b f = \int_a^b f\).

d) Find an example of an \(f\) with a singularity at \(c\) as above such that \(p.v.\!\int_a^b f\) exists, but the improper integrals \(\int_a^c f\) and \(\int_c^b f\) diverge.

e) Suppose \(f \colon [-1,1] \to {\mathbb{R}}\) is continuous. Show that \(p.v.\!\int_{-1}^1 \frac{f(x)}{x}~dx\) exists.

Let \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) and \(g \colon {\mathbb{R}}\to {\mathbb{R}}\) be continuous functions, where \(g(x) = 0\) for all \(x \notin [a,b]\) for some interval \([a,b]\).

a) Show that the *convolution* \[(g * f)(x) := \int_{-\infty}^\infty f(t)g(x-t)~dt\] is well-defined for all \(x \in {\mathbb{R}}\).

b) Suppose \(\int_{-\infty}^\infty \left\lvert {f(x)} \right\rvert~dx < \infty\). Prove that \[\lim_{x \to -\infty} (g * f)(x) = 0, \qquad \text{and} \qquad \lim_{x \to \infty} (g * f)(x) = 0 .\]

## Sequences of Functions

### Pointwise and uniform convergence

*Note: 1–1.5 lecture*

Up till now when we talked about sequences we always talked about sequences of numbers. However, a very useful concept in analysis is to use a sequence of functions. For example, a solution to some differential equation might be found by finding only approximate solutions. Then the real solution is some sort of limit of those approximate solutions.

When talking about sequences of functions, the tricky part is that there are multiple notions of a limit. Let us describe two common notions of a limit of a sequence of functions.

#### Pointwise convergence

For every \(n \in {\mathbb{N}}\) let \(f_n \colon S \to {\mathbb{R}}\) be a function. We say the sequence \(\{ f_n \}_{n=1}^\infty\) *converges pointwise* to \(f \colon S \to {\mathbb{R}}\), if for every \(x \in S\) we have \[f(x) = \lim_{n\to\infty} f_n(x) .\]

It is common to say that \(f_n \colon S \to {\mathbb{R}}\) *converges to \(f\) on \(T \subset {\mathbb{R}}\)* for some \(f \colon T \to {\mathbb{R}}\). In that case we, of course, mean \(f(x) = \lim\, f_n(x)\) for every \(x \in T\). We simply mean that the restrictions of \(f_n\) to \(T\) converge pointwise to \(f\).

The sequence of functions defined by \(f_n(x) := x^{2n}\) converges to \(f \colon [-1,1] \to {\mathbb{R}}\) on \([-1,1]\), where \[f(x) = \begin{cases} 1 & \text{if $x=-1$ or $x=1$,} \\ 0 & \text{otherwise.} \end{cases}\] See .

To see this is so, first take \(x \in (-1,1)\). Then \(0 \leq x^2 < 1\). We have seen before that \[\left\lvert {x^{2n} - 0} \right\rvert = {(x^2)}^n \to 0 \quad \text{as} \quad n \to \infty .\] Therefore \(\lim\,f_n(x) = 0\).

When \(x = 1\) or \(x=-1\), then \(x^{2n} = 1\) for all \(n\) and hence \(\lim\,f_n(x) = 1\). We also note that \(\{ f_n(x) \}\) does not converge for all other \(x\).

Often, functions are given as a series. In this case, we use the notion of pointwise convergence to find the values of the function.

We write \[\sum_{k=0}^\infty x^k\] to denote the limit of the functions \[f_n(x) := \sum_{k=0}^n x^k .\] When studying series, we have seen that on \(x \in (-1,1)\) the \(f_n\) converge pointwise to \[\frac{1}{1-x} .\]

The subtle point here is that while \(\frac{1}{1-x}\) is defined for all \(x \not=1\), and \(f_n\) are defined for all \(x\) (even at \(x=1\)), convergence only happens on \((-1,1)\).

Therefore, when we write \[f(x) := \sum_{k=0}^\infty x^k\] we mean that \(f\) is defined on \((-1,1)\) and is the pointwise limit of the partial sums.

Let \(f_n(x) := \sin(xn)\). Then \(f_n\) does not converge pointwise to any function on any interval. It may converge at certain points, such as when \(x=0\) or \(x=\pi\). It is left as an exercise that in any interval \([a,b]\), there exists an \(x\) such that \(\sin(xn)\) does not have a limit as \(n\) goes to infinity.

Before we move to uniform convergence, let us reformulate pointwise convergence in a different way. We leave the proof to the reader, it is a simple application of the definition of convergence of a sequence of real numbers.

[ptwsconv:prop] Let \(f_n \colon S \to {\mathbb{R}}\) and \(f \colon S \to {\mathbb{R}}\) be functions. Then \(\{ f_n \}\) converges pointwise to \(f\) if and only if for every \(x \in S\), and every \(\epsilon > 0\), there exists an \(N \in {\mathbb{N}}\) such that \[\left\lvert {f_n(x)-f(x)} \right\rvert < \epsilon\] for all \(n \geq N\).

The key point here is that \(N\) can depend on \(x\), not just on \(\epsilon\). That is, for each \(x\) we can pick a different \(N\). If we can pick one \(N\) for all \(x\), we have what is called uniform convergence.

#### Uniform convergence

Let \(f_n \colon S \to {\mathbb{R}}\) be functions. We say the sequence \(\{ f_n \}\) *converges uniformly* to \(f \colon S \to {\mathbb{R}}\), if for every \(\epsilon > 0\) there exists an \(N \in {\mathbb{N}}\) such that for all \(n \geq N\) we have \[\left\lvert {f_n(x) - f(x)} \right\rvert < \epsilon \qquad \text{for all $x \in S$.}\]

Note that \(N\) now cannot depend on \(x\). Given \(\epsilon > 0\) we must find an \(N\) that works for all \(x \in S\). Because of , we see that uniform convergence implies pointwise convergence.

Let \(\{ f_n \}\) be a sequence of functions \(f_n \colon S \to {\mathbb{R}}\). If \(\{ f_n \}\) converges uniformly to \(f \colon S \to {\mathbb{R}}\), then \(\{ f_n \}\) converges pointwise to \(f\).

The converse does not hold.

The functions \(f_n(x) := x^{2n}\) do not converge uniformly on \([-1,1]\), even though they converge pointwise. To see this, suppose for contradiction that the convergence is uniform. For \(\epsilon := \nicefrac{1}{2}\), there would have to exist an \(N\) such that \(x^{2N} = \left\lvert {x^{2N} - 0} \right\rvert < \nicefrac{1}{2}\) for all \(x \in (-1,1)\) (as \(f_n(x)\) converges to 0 on \((-1,1)\)). But that means that for any sequence \(\{ x_k \}\) in \((-1,1)\) such that \(\lim\, x_k = 1\) we have \(x_k^{2N} < \nicefrac{1}{2}\) for all \(k\). On the other hand \(x^{2N}\) is a continuous function of \(x\) (it is a polynomial), therefore we obtain a contradiction \[1 = 1^{2N} = \lim_{k\to\infty} x_k^{2N} \leq \nicefrac{1}{2} .\]

However, if we restrict our domain to \([-a,a]\) where \(0 < a < 1\), then \(\{ f_n \}\) converges uniformly to 0 on \([-a,a]\). First note that \(a^{2n} \to 0\) as \(n \to \infty\). Thus given \(\epsilon > 0\), pick \(N \in {\mathbb{N}}\) such that \(a^{2n} < \epsilon\) for all \(n \geq N\). Then for any \(x \in [-a,a]\) we have \(\left\lvert {x} \right\rvert \leq a\). Therefore, for \(n \geq N\) \[\left\lvert {x^{2N}} \right\rvert = \left\lvert {x} \right\rvert^{2N} \leq a^{2N} < \epsilon .\]

#### Convergence in uniform norm

For bounded functions there is another more abstract way to think of uniform convergence. To every bounded function we assign a certain nonnegative number (called the uniform norm). This number measures the “distance” of the function from 0. We can then “measure” how far two functions are from each other. We simply translate a statement about uniform convergence into a statement about a certain sequence of real numbers converging to zero.

[def:unifnorm] Let \(f \colon S \to {\mathbb{R}}\) be a bounded function. Define \[\left\lVert {f} \right\rVert_u := \sup \bigl\{ \left\lvert {f(x)} \right\rvert : x \in S \bigr\} .\] \(\left\lVert {\cdot} \right\rVert_u\) is called the *uniform norm*.

To use this notation ^{26} and this concept, the domain \(S\) must be fixed. Some authors use the notation \(\left\lVert {f} \right\rVert_S\) to emphasize the dependence on \(S\).

A sequence of bounded functions \(f_n \colon S \to {\mathbb{R}}\) converges uniformly to \(f \colon S \to {\mathbb{R}}\), if and only if \[\lim_{n\to\infty} \left\lVert {f_n - f} \right\rVert_u = 0 .\]

First suppose \(\lim \left\lVert {f_n - f} \right\rVert_u = 0\). Let \(\epsilon > 0\) be given. Then there exists an \(N\) such that for \(n \geq N\) we have \(\left\lVert {f_n - f} \right\rVert_u < \epsilon\). As \(\left\lVert {f_n-f} \right\rVert_u\) is the supremum of \(\left\lvert {f_n(x)-f(x)} \right\rvert\), we see that for all \(x\) we have \(\left\lvert {f_n(x)-f(x)} \right\rvert < \epsilon\).

On the other hand, suppose \(\{ f_n \}\) converges uniformly to \(f\). Let \(\epsilon > 0\) be given. Then find \(N\) such that \(\left\lvert {f_n(x)-f(x)} \right\rvert < \epsilon\) for all \(x \in S\). Taking the supremum we see that \(\left\lVert {f_n - f} \right\rVert_u < \epsilon\). Hence \(\lim \left\lVert {f_n-f} \right\rVert_u = 0\).

Sometimes it is said that *\(\{ f_n \}\) converges to \(f\) in uniform norm* instead of *converges uniformly*. The proposition says that the two notions are the same thing.

Let \(f_n \colon [0,1] \to {\mathbb{R}}\) be defined by \(f_n(x) := \frac{nx+ \sin(nx^2)}{n}\). Then we claim \(\{ f_n \}\) converges uniformly to \(f(x) := x\). Let us compute: \[\begin{split} \left\lVert {f_n-f} \right\rVert_u & = \sup \left\{ \left\lvert {\frac{nx+ \sin(nx^2)}{n} - x} \right\rvert : x \in [0,1] \right\} \\ & = \sup \left\{ \frac{\left\lvert {\sin(nx^2)} \right\rvert}{n} : x \in [0,1] \right\} \\ & \leq \sup \{ \nicefrac{1}{n} : x \in [0,1] \} \\ & = \nicefrac{1}{n}. \end{split}\]

Using uniform norm, we define Cauchy sequences in a similar way as we define Cauchy sequences of real numbers.

Let \(f_n \colon S \to {\mathbb{R}}\) be bounded functions. The sequence is *Cauchy in the uniform norm* or *uniformly Cauchy* if for every \(\epsilon > 0\), there exists an \(N \in {\mathbb{N}}\) such that for \(m,k \geq N\) we have \[\left\lVert {f_m-f_k} \right\rVert_u < \epsilon .\]

[prop:uniformcauchy] Let \(f_n \colon S \to {\mathbb{R}}\) be bounded functions. Then \(\{ f_n \}\) is Cauchy in the uniform norm if and only if there exists an \(f \colon S \to {\mathbb{R}}\) and \(\{ f_n \}\) converges uniformly to \(f\).

Let us first suppose \(\{ f_n \}\) is Cauchy in the uniform norm. Let us define \(f\). Fix \(x\), then the sequence \(\{ f_n(x) \}\) is Cauchy because \[\left\lvert {f_m(x)-f_k(x)} \right\rvert \leq \left\lVert {f_m-f_k} \right\rVert_u .\] Thus \(\{ f_n(x) \}\) converges to some real number. Define \(f \colon S \to {\mathbb{R}}\) by \[f(x) := \lim_{n \to \infty} f_n(x) .\] The sequence \(\{ f_n \}\) converges pointwise to \(f\). To show that the convergence is uniform, let \(\epsilon > 0\) be given. Find an \(N\) such that for \(m, k \geq N\) we have \(\left\lVert {f_m-f_k} \right\rVert_u < \nicefrac{\epsilon}{2}\). In other words for all \(x\) we have \(\left\lvert {f_m(x)-f_k(x)} \right\rvert < \nicefrac{\epsilon}{2}\). We take the limit as \(k\) goes to infinity. Then \(\left\lvert {f_m(x)-f_k(x)} \right\rvert\) goes to \(\left\lvert {f_m(x)-f(x)} \right\rvert\). Consequently for all \(x\) we get \[\left\lvert {f_m(x)-f(x)} \right\rvert \leq \nicefrac{\epsilon}{2} < \epsilon .\] And hence \(\{ f_n \}\) converges uniformly.

For the other direction, suppose \(\{ f_n \}\) converges uniformly to \(f\). Given \(\epsilon > 0\), find \(N\) such that for all \(n \geq N\) we have \(\left\lvert {f_n(x)-f(x)} \right\rvert < \nicefrac{\epsilon}{4}\) for all \(x \in S\). Therefore for all \(m, k \geq N\) we have \[\left\lvert {f_m(x)-f_k(x)} \right\rvert = \left\lvert {f_m(x)-f(x)+f(x)-f_k(x)} \right\rvert \leq \left\lvert {f_m(x)-f(x)} \right\rvert+\left\lvert {f(x)-f_k(x)} \right\rvert < \nicefrac{\epsilon}{4} + \nicefrac{\epsilon}{4} .\] Take supremum over all \(x\) to obtain \[\left\lVert {f_m-f_k} \right\rVert_u \leq \nicefrac{\epsilon}{2} < \epsilon . \qedhere\]

## Exercises

Let \(f\) and \(g\) be bounded functions on \([a,b]\). Prove \[\left\lVert {f+g} \right\rVert_u \leq \left\lVert {f} \right\rVert_u + \left\lVert {g} \right\rVert_u .\]

a) Find the pointwise limit \(\dfrac{e^{x/n}}{n}\) for \(x \in {\mathbb{R}}\).

b) Is the limit uniform on \({\mathbb{R}}\)?

c) Is the limit uniform on \([0,1]\)?

Suppose \(f_n \colon S \to {\mathbb{R}}\) are functions that converge uniformly to \(f \colon S \to {\mathbb{R}}\). Suppose \(A \subset S\). Show that the sequence of restrictions \(\{ f_n|_A \}\) converges uniformly to \(f|_A\).

Suppose \(\{ f_n \}\) and \(\{ g_n \}\) defined on some set \(A\) converge to \(f\) and \(g\) respectively pointwise. Show that \(\{ f_n+g_n \}\) converges pointwise to \(f+g\).

Suppose \(\{ f_n \}\) and \(\{ g_n \}\) defined on some set \(A\) converge to \(f\) and \(g\) respectively uniformly on \(A\). Show that \(\{ f_n+g_n \}\) converges uniformly to \(f+g\) on \(A\).

Find an example of a sequence of functions \(\{ f_n \}\) and \(\{ g_n \}\) that converge uniformly to some \(f\) and \(g\) on some set \(A\), but such that \(\{ f_ng_n \}\) (the multiple) does not converge uniformly to \(fg\) on \(A\). Hint: Let \(A := {\mathbb{R}}\), let \(f(x):=g(x) := x\). You can even pick \(f_n = g_n\).

Suppose there exists a sequence of functions \(\{ g_n \}\) uniformly converging to \(0\) on \(A\). Now suppose we have a sequence of functions \(\{ f_n \}\) and a function \(f\) on \(A\) such that \[\left\lvert {f_n(x) - f(x)} \right\rvert \leq g_n(x)\] for all \(x \in A\). Show that \(\{ f_n \}\) converges uniformly to \(f\) on \(A\).

Let \(\{ f_n \}\), \(\{ g_n \}\) and \(\{ h_n \}\) be sequences of functions on \([a,b]\). Suppose \(\{ f_n \}\) and \(\{ h_n \}\) converge uniformly to some function \(f \colon [a,b] \to {\mathbb{R}}\) and suppose \(f_n(x) \leq g_n(x) \leq h_n(x)\) for all \(x \in [a,b]\). Show that \(\{ g_n \}\) converges uniformly to \(f\).

Let \(f_n \colon [0,1] \to {\mathbb{R}}\) be a sequence of increasing functions (that is, \(f_n(x) \geq f_n(y)\) whenever \(x \geq y\)). Suppose \(f_n(0) = 0\) and \(\lim\limits_{n \to \infty} f_n(1) = 0\). Show that \(\{ f_n \}\) converges uniformly to \(0\).

Let \(\{f_n\}\) be a sequence of functions defined on \([0,1]\). Suppose there exists a sequence of distinct numbers \(x_n \in [0,1]\) such that \[f_n(x_n) = 1 .\] Prove or disprove the following statements:

a) True or false: There exists \(\{ f_n \}\) as above that converges to \(0\) pointwise.

b) True or false: There exists \(\{ f_n \}\) as above that converges to \(0\) uniformly on \([0,1]\).

Fix a continuous \(h \colon [a,b] \to {\mathbb{R}}\). Let \(f(x) := h(x)\) for \(x \in [a,b]\), \(f(x) := h(a)\) for \(x < a\) and \(f(x) := h(b)\) for all \(x > b\). First show that \(f \colon {\mathbb{R}}\to {\mathbb{R}}\) is continuous. Now let \(f_n\) be the function \(g\) from with \(\epsilon = \nicefrac{1}{n}\), defined on the interval \([a,b]\). Show that \(\{ f_n \}\) converges uniformly to \(h\) on \([a,b]\).

### Interchange of limits

*Note: 1–1.5 lectures*

Large parts of modern analysis deal mainly with the question of the interchange of two limiting operations. When we have a chain of two limits, we cannot always just swap the limits. For example, \[0 = \lim_{n\to\infty} \left( \lim_{k\to\infty} \frac{\nicefrac{n}{k}}{\nicefrac{n}{k} + 1} \right) \not= \lim_{k\to\infty} \left( \lim_{n\to\infty} \frac{\nicefrac{n}{k}}{\nicefrac{n}{k} + 1} \right) = 1 .\]

When talking about sequences of functions, interchange of limits comes up quite often. We treat two cases. First we look at continuity of the limit, and second we look at the integral of the limit.

#### Continuity of the limit

If we have a sequence \(\{ f_n \}\) of continuous functions, is the limit continuous? Suppose \(f\) is the (pointwise) limit of \(\{ f_n \}\). If \(\lim\, x_k = x\) we are interested in the following interchange of limits. The equality we have to prove (it is not always true) is marked with a question mark. In fact the limits to the left of the question mark might not even exist. \[\lim_{k \to \infty} f(x_k) = \lim_{k \to \infty} \Bigl( \lim_{n \to \infty} f_n(x_k) \Bigr) \overset{\text{\textbf{?}}}{=} \lim_{n \to \infty} \Bigl( \lim_{k \to \infty} f_n(x_k) \Bigr) = \lim_{n \to \infty} f_n(x) = f(x) .\] In particular, we wish to find conditions on the sequence \(\{ f_n \}\) so that the above equation holds. It turns out that if we only require pointwise convergence, then the limit of a sequence of functions need not be continuous, and the above equation need not hold.

Let \(f_n \colon [0,1] \to {\mathbb{R}}\) be defined as \[f_n(x) := \begin{cases} 1-nx & \text{if $x < \nicefrac{1}{n}$,}\\ 0 & \text{if $x \geq \nicefrac{1}{n}$.} \end{cases}\] See .

Each function \(f_n\) is continuous. Fix an \(x \in (0,1]\). If \(n \geq \nicefrac{1}{x}\), then \(x \geq \nicefrac{1}{n}\). Therefore for \(n \geq \nicefrac{1}{x}\) we have \(f_n(x) = 0\), and so \[\lim_{n \to \infty} f_n(x) = 0.\] On the other hand if \(x=0\), then \[\lim_{n \to \infty} f_n(0) = \lim_{n \to \infty} 1 = 1.\] Thus the pointwise limit of \(f_n\) is the function \(f \colon [0,1] \to {\mathbb{R}}\) defined by \[f(x) := \begin{cases} 1 & \text{if $x = 0$,}\\ 0 & \text{if $x > 0$.} \end{cases}\] The function \(f\) is not continuous at 0.

If we, however, require the convergence to be uniform, the limits can be interchanged.

Let \(\{ f_n \}\) be a sequence of continuous functions \(f_n \colon S \to {\mathbb{R}}\) converging uniformly to \(f \colon S \to {\mathbb{R}}\). Then \(f\) is continuous.

Let \(x \in S\) be fixed. Let \(\{ x_n \}\) be a sequence in \(S\) converging to \(x\).

Let \(\epsilon > 0\) be given. As \(\{ f_k \}\) converges uniformly to \(f\), we find a \(k \in {\mathbb{N}}\) such that \[\left\lvert {f_k(y)-f(y)} \right\rvert < \nicefrac{\epsilon}{3}\] for all \(y \in S\). As \(f_k\) is continuous at \(x\), we find an \(N \in {\mathbb{N}}\) such that for \(m \geq N\) we have \[\left\lvert {f_k(x_m)-f_k(x)} \right\rvert < \nicefrac{\epsilon}{3} .\] Thus for \(m \geq N\) we have \[\begin{split} \left\lvert {f(x_m)-f(x)} \right\rvert & = \left\lvert {f(x_m)-f_k(x_m)+f_k(x_m)-f_k(x)+f_k(x)-f(x)} \right\rvert \\ & \leq \left\lvert {f(x_m)-f_k(x_m)} \right\rvert+ \left\lvert {f_k(x_m)-f_k(x)} \right\rvert+ \left\lvert {f_k(x)-f(x)} \right\rvert \\ & < \nicefrac{\epsilon}{3} + \nicefrac{\epsilon}{3} + \nicefrac{\epsilon}{3} = \epsilon . \end{split}\] Therefore \(\{ f(x_m) \}\) converges to \(f(x)\) and hence \(f\) is continuous at \(x\). As \(x\) was arbitrary, \(f\) is continuous everywhere.

#### Integral of the limit

Again, if we simply require pointwise convergence, then the integral of a limit of a sequence of functions need not be equal to the limit of the integrals.

Let \(f_n \colon [0,1] \to {\mathbb{R}}\) be defined as \[f_n(x) := \begin{cases} 0 & \text{if $x = 0$,}\\ n-n^2x & \text{if $0 < x < \nicefrac{1}{n}$,}\\ 0 & \text{if $x \geq \nicefrac{1}{n}$.} \end{cases}\] See .

Each \(f_n\) is Riemann integrable (it is continuous on \((0,1]\) and bounded), and it is easy to see \[\int_0^1 f_n = \int_0^{\nicefrac{1}{n}} (n-n^2x)~dx = \nicefrac{1}{2} .\] Let us compute the pointwise limit of \(\{ f_n \}\). Fix an \(x \in (0,1]\). For \(n \geq \nicefrac{1}{x}\) we have \(x \geq \nicefrac{1}{n}\) and so \(f_n(x) = 0\). Therefore \[\lim_{n \to \infty} f_n(x) = 0.\] We also have \(f_n(0) = 0\) for all \(n\). Therefore the pointwise limit of \(\{ f_n \}\) is the zero function. Thus \[\nicefrac{1}{2} = \lim_{n\to\infty} \int_0^1 f_n (x)~dx \not= \int_0^1 \left( \lim_{n\to\infty} f_n(x)\right)~dx = \int_0^1 0~dx = 0 .\]

But if we again require the convergence to be uniform, the limits can be interchanged.

[integralinterchange:thm] Let \(\{ f_n \}\) be a sequence of Riemann integrable functions \(f_n \colon [a,b] \to {\mathbb{R}}\) converging uniformly to \(f \colon [a,b] \to {\mathbb{R}}\). Then \(f\) is Riemann integrable and \[\int_a^b f = \lim_{n\to\infty} \int_a^b f_n .\]

Let \(\epsilon > 0\) be given. As \(f_n\) goes to \(f\) uniformly, we find an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \(\left\lvert {f_n(x)-f(x)} \right\rvert < \frac{\epsilon}{2(b-a)}\) for all \(x \in [a,b]\). In particular, by reverse triangle inequality \(\left\lvert {f(x)} \right\rvert < \frac{\epsilon}{2(b-a)} + \left\lvert {f_n(x)} \right\rvert\) for all \(x\), hence \(f\) is bounded as \(f_n\) is bounded. Note that \(f_n\) is integrable and compute \[\begin{split} \overline{\int_a^b} f - \underline{\int_a^b} f & = \overline{\int_a^b} \bigl( f(x) - f_n(x) + f_n(x) \bigr)~dx - \underline{\int_a^b} \bigl( f(x) - f_n(x) + f_n(x) \bigr)~dx \\ & \leq \overline{\int_a^b} \bigl( f(x) - f_n(x) \bigr)~dx + \overline{\int_a^b} f_n(x) ~dx - \underline{\int_a^b} \bigl( f(x) - f_n(x) \bigr)~dx - \underline{\int_a^b} f_n(x) ~dx \\ & = \overline{\int_a^b} \bigl( f(x) - f_n(x) \bigr)~dx + \int_a^b f_n(x) ~dx - \underline{\int_a^b} \bigl( f(x) - f_n(x) \bigr)~dx - \int_a^b f_n(x) ~dx \\ & = \overline{\int_a^b} \bigl( f(x) - f_n(x) \bigr)~dx - \underline{\int_a^b} \bigl( f(x) - f_n(x) \bigr)~dx \\ & \leq \frac{\epsilon}{2(b-a)} (b-a) + \frac{\epsilon}{2(b-a)} (b-a) = \epsilon . \end{split}\] The first inequality is (it follows as supremum of a sum is less than or equal to the sum of suprema and similarly for infima, see ). The second inequality follows from and the fact that for all \(x \in [a,b]\) we have \(\frac{-\epsilon}{2(b-a)} < f(x)-f_n(x) < \frac{\epsilon}{2(b-a)}\). As \(\epsilon > 0\) was arbitrary, \(f\) is Riemann integrable.

Finally we compute \(\int_a^b f\). We apply in the calculation. Again, for \(n \geq M\) (where \(M\) is the same as above) we have \[\begin{split} \left\lvert {\int_a^b f - \int_a^b f_n} \right\rvert & = \left\lvert { \int_a^b \bigl(f(x) - f_n(x)\bigr)~dx} \right\rvert \\ & \leq \frac{\epsilon}{2(b-a)} (b-a) = \frac{\epsilon}{2} < \epsilon . \end{split}\] Therefore \(\{ \int_a^b f_n \}\) converges to \(\int_a^b f\).

Suppose we wish to compute \[\lim_{n\to\infty} \int_0^1 \frac{nx+ \sin(nx^2)}{n} ~dx .\] It is impossible to compute the integrals for any particular \(n\) using calculus as \(\sin(nx^2)\) has no closed-form antiderivative. However, we can compute the limit. We have shown before that \(\frac{nx+ \sin(nx^2)}{n}\) converges uniformly on \([0,1]\) to \(x\). By , the limit exists and \[\lim_{n\to\infty} \int_0^1 \frac{nx+ \sin(nx^2)}{n} ~dx = \int_0^1 x ~dx = \nicefrac{1}{2} .\]

If convergence is only pointwise, the limit need not even be Riemann integrable. On \([0,1]\) define \[f_n(x) := \begin{cases} 1 & \text{if $x = \nicefrac{p}{q}$ in lowest terms and $q \leq n$,} \\ 0 & \text{otherwise.} \end{cases}\] The function \(f_n\) differs from the zero function at finitely many points; there are only finitely many fractions in \([0,1]\) with denominator less than or equal to \(n\). So \(f_n\) is integrable and \(\int_0^1 f_n = \int_0^1 0 = 0\). It is an easy exercise to show that \(\{ f_n \}\) converges pointwise to the Dirichlet function \[f(x) := \begin{cases} 1 & \text{if $x \in {\mathbb{Q}}$,} \\ 0 & \text{otherwise,} \end{cases}\] which is not Riemann integrable.

In fact, if the convergence is only pointwise, the limit of bounded functions is not even necessarily bounded. Define \(f_n \colon [0,1] \to {\mathbb{R}}\) by \[f_n(x) := \begin{cases} 0 & \text{ if $x < \nicefrac{1}{n}$,}\\ \nicefrac{1}{x} & \text{ else.} \end{cases}\] For every \(n\) we get that \(\left\lvert {f_n(x)} \right\rvert \leq n\) for all \(x \in [0,1]\) so the functions are bounded. However \(f_n\) converge pointwise to \[f(x) := \begin{cases} 0 & \text{ if $x = 0$,}\\ \nicefrac{1}{x} & \text{ else,} \end{cases}\] which is unbounded.

Let us remark that while uniform convergence is enough to swap limits with integrals, it is not, however, enough to swap limits with derivatives, unless you also have uniform convergence of the derivatives themselves. See the exercises below.

## Exercises

While uniform convergence preserves continuity, it does not preserve differentiability. Find an explicit example of a sequence of differentiable functions on \([-1,1]\) that converge uniformly to a function \(f\) such that \(f\) is not differentiable. Hint: Consider \(\left\lvert {x} \right\rvert^{1+1/n}\), show that these functions are differentiable, converge uniformly, and then show that the limit is not differentiable.

Let \(f_n(x) = \frac{x^n}{n}\). Show that \(\{ f_n \}\) converges uniformly to a differentiable function \(f\) on \([0,1]\) (find \(f\)). However, show that \(f'(1) \not= \lim\limits_{n\to\infty} f_n'(1)\).

Note: The previous two exercises show that we cannot simply swap limits with derivatives, even if the convergence is uniform. See also below.

Let \(f \colon [0,1] \to {\mathbb{R}}\) be a Riemann integrable (hence bounded) function. Find \(\displaystyle \lim_{n\to\infty} \int_0^1 \frac{f(x)}{n} ~dx\).

Show \(\displaystyle \lim_{n\to\infty} \int_1^2 e^{-nx^2} ~dx = 0\). Feel free to use what you know about the exponential function from calculus.

Find an example of a sequence of continuous functions on \((0,1)\) that converges pointwise to a continuous function on \((0,1)\), but the convergence is not uniform.

Note: In the previous exercise, \((0,1)\) was picked for simplicity. For a more challenging exercise, replace \((0,1)\) with \([0,1]\).

True/False; prove or find a counterexample to the following statement: If \(\{ f_n \}\) is a sequence of everywhere discontinuous functions on \([0,1]\) that converge uniformly to a function \(f\), then \(f\) is everywhere discontinuous.

[c1uniflim:exercise] For a continuously differentiable function \(f \colon [a,b] \to {\mathbb{R}}\), define \[\left\lVert {f} \right\rVert_{C^1} := \left\lVert {f} \right\rVert_u + \left\lVert {f'} \right\rVert_u .\] Suppose \(\{ f_n \}\) is a sequence of continuously differentiable functions such that for every \(\epsilon >0\), there exists an \(M\) such that for all \(n,k \geq M\) we have \[\left\lVert {f_n-f_k} \right\rVert_{C^1} < \epsilon .\] Show that \(\{ f_n \}\) converges uniformly to some continuously differentiable function \(f \colon [a,b] \to {\mathbb{R}}\).

For the following two exercises let us define for a Riemann integrable function \(f \colon [0,1] \to {\mathbb{R}}\) the following number \[\left\lVert {f} \right\rVert_{L^1} := \int_0^1 \left\lvert {f(x)} \right\rvert~dx .\] It is true that \(\left\lvert {f} \right\rvert\) is integrable whenever \(f\) is, see . This norm defines another very common type of convergence called the \(L^1\)-convergence, that is however a bit more subtle.

Suppose \(\{ f_n \}\) is a sequence of Riemann integrable functions on \([0,1]\) that converges uniformly to \(0\). Show that \[\lim_{n\to\infty} \left\lVert {f_n} \right\rVert_{L^1} = 0 .\]

Find a sequence of Riemann integrable functions \(\{ f_n \}\) on \([0,1]\) that converges pointwise to \(0\), but \[\lim_{n\to\infty} \left\lVert {f_n} \right\rVert_{L^1} \text{ does not exist (is $\infty$).}\]

Prove *Dini’s theorem*: Let \(f_n \colon [a,b] \to {\mathbb{R}}\) be a sequence of continuous functions such that \[0 \leq f_{n+1}(x) \leq f_n(x) \leq \cdots \leq f_1(x) \qquad \text{for all $n \in {\mathbb{N}}$.}\] Suppose \(\{ f_n \}\) converges pointwise to \(0\). Show that \(\{ f_n \}\) converges to zero uniformly.

Suppose \(f_n \colon [a,b] \to {\mathbb{R}}\) is a sequence of continuous functions that converges pointwise to a continuous \(f \colon [a,b] \to {\mathbb{R}}\). Suppose that for any \(x \in [a,b]\) the sequence \(\{ \left\lvert {f_n(x)-f(x)} \right\rvert \}\) is monotone. Show that the sequence \(\{f_n\}\) converges uniformly.

Find a sequence of Riemann integrable functions \(f_n \colon [0,1] \to {\mathbb{R}}\) such that \(\{ f_n \}\) converges to zero pointwise, and such that a) \(\bigl\{ \int_0^1 f_n \bigr\}_{n=1}^\infty\) increases without bound, b) \(\bigl\{ \int_0^1 f_n \bigr\}_{n=1}^\infty\) is the sequence \(-1,1,-1,1,-1,1, \ldots\).

It is possible to define a *joint limit* of a double sequence \(\{ x_{n,m} \}\) of real numbers (that is a function from \({\mathbb{N}}\times {\mathbb{N}}\) to \({\mathbb{R}}\)). We say \(L\) is the joint limit of \(\{ x_{n,m} \}\) and write \[\lim_{\substack{n\to\infty\\m\to\infty}} x_{n,m} = L , \qquad \text{or} \qquad \lim_{(n,m) \to \infty} x_{n,m} = L ,\] if for every \(\epsilon > 0\), there exists an \(M\) such that if \(n \geq M\) and \(m \geq M\), then \(\left\lvert {x_{n,m} - L} \right\rvert < \epsilon\).

Suppose the joint limit of \(\{ x_{n,m} \}\) is \(L\), and suppose that for all \(n\), \(\lim\limits_{m \to \infty} x_{n,m}\) exists, and for all \(m\), \(\lim\limits_{n \to \infty} x_{n,m}\) exists. Then show \(\lim\limits_{n\to\infty}\lim\limits_{m \to \infty} x_{n,m} = \lim\limits_{m\to\infty}\lim\limits_{n \to \infty} x_{n,m} = L\).

A joint limit does not mean the iterated limits even exist. Consider \(x_{n,m} := \frac

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[36]/p[18]/span, line 1, column 5 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

a) Show that for no \(n\) does \(\lim\limits_{m \to \infty} x_{n,m}\) exist, and for no \(m\) does \(\lim\limits_{n \to \infty} x_{n,m}\) exist. So neither \(\lim\limits_{n\to\infty}\lim\limits_{m \to \infty} x_{n,m}\) nor \(\lim\limits_{m\to\infty}\lim\limits_{n \to \infty} x_{n,m}\) makes any sense at all.

b) Show that the joint limit of \(\{ x_{n,m} \}\) exists and is 0.

### Picard’s theorem

*Note: 1–2 lectures (can be safely skipped)*

A first semester course in analysis should have a *pièce de résistance* caliber theorem. We pick a theorem whose proof combines everything we have learned. It is more sophisticated than the fundamental theorem of calculus, the first highlight theorem of this course. The theorem we are talking about is Picard’s theorem ^{27} on existence and uniqueness of a solution to an ordinary differential equation. Both the statement and the proof are beautiful examples of what one can do with all we have learned. It is also a good example of how analysis is applied as differential equations are indispensable in science.

#### First order ordinary differential equation

Modern science is described in the language of *differential equations*. That is, equations involving not only the unknown, but also its derivatives. The simplest nontrivial form of a differential equation is the so-called *first order ordinary differential equation* \[y' = F(x,y) .\] Generally we also specify \(y(x_0)=y_0\). The solution of the equation is a function \(y(x)\) such that \(y(x_0)=y_0\) and \(y'(x) = F\bigl(x,y(x)\bigr)\).

When \(F\) involves only the \(x\) variable, the solution is given by the fundamental theorem of calculus. On the other hand, when \(F\) depends on both \(x\) and \(y\) we need far more firepower. It is not always true that a solution exists, and if it does, that it is the unique solution. Picard’s theorem gives us certain sufficient conditions for existence and uniqueness.

#### The theorem

We need a definition of continuity in two variables. First, a point in the plane \({\mathbb{R}}^2 = {\mathbb{R}}\times {\mathbb{R}}\) is denoted by an ordered pair \((x,y)\). To make matters simple, let us give the following sequential definition of continuity.

Let \(U \subset {\mathbb{R}}^2\) be a set and \(F \colon U \to {\mathbb{R}}\) be a function. Let \((x,y) \in U\) be a point. The function \(F\) is *continuous* at \((x,y)\) if for every sequence \(\{ (x_n,y_n) \}_{n=1}^\infty\) of points in \(U\) such that \(\lim\, x_n = x\) and \(\lim\, y_n = y\), we have \[\lim_{n \to \infty} F(x_n,y_n) = F(x,y) .\] We say \(F\) is continuous if it is continuous at all points in \(U\).

Let \(I, J \subset {\mathbb{R}}\) be closed bounded intervals, let \(I_0\) and \(J_0\) be their interiors, and let \((x_0,y_0) \in I_0 \times J_0\). Suppose \(F \colon I \times J \to {\mathbb{R}}\) is continuous and Lipschitz in the second variable, that is, there exists a number \(L\) such that \[\left\lvert {F(x,y) - F(x,z)} \right\rvert \leq L \left\lvert {y-z} \right\rvert \ \ \ \text{ for all $y,z \in J$, $x \in I$} .\] Then there exists an \(h > 0\) and a unique differentiable function \(f \colon [x_0 - h, x_0 + h] \to J \subset {\mathbb{R}}\), such that \[\label{picard:diffeq} f'(x) = F\bigl(x,f(x)\bigr) \qquad \text{and} \qquad f(x_0) = y_0.\]

Suppose we could find a solution \(f\). Using the fundamental theorem of calculus we integrate the equation \(f'(x) = F\bigl(x,f(x)\bigr)\), \(f(x_0) = y_0\), and write [picard:diffeq] as the integral equation \[\label{picard:inteq} f(x) = y_0 + \int_{x_0}^x F\bigl(t,f(t)\bigr)~dt .\] The idea of our proof is that we try to plug in approximations to a solution to the right-hand side of [picard:inteq] to get better approximations on the left hand side of [picard:inteq]. We hope that in the end the sequence converges and solves [picard:inteq] and hence [picard:diffeq]. The technique below is called *Picard iteration*, and the individual functions \(f_k\) are called the *Picard iterates*.

Without loss of generality, suppose \(x_0 = 0\) (exercise below). Another exercise tells us that \(F\) is bounded as it is continuous. Therefore pick some \(M > 0\) so that \(\left\lvert {F(x,y)} \right\rvert \leq M\) for all \((x,y) \in I\times J\). Pick \(\alpha > 0\) such that \([-\alpha,\alpha] \subset I\) and \([y_0-\alpha, y_0 + \alpha] \subset J\). Define \[h := \min \left\{ \alpha, \frac{\alpha}{M+L\alpha} \right\} .\] Observe \([-h,h] \subset I\).

Set \(f_0(x) := y_0\). We define \(f_k\) inductively. Assuming \(f_{k-1}([-h,h]) \subset [y_0-\alpha,y_0+\alpha]\), we see \(F\bigl(t,f_{k-1}(t)\bigr)\) is a well defined function of \(t\) for \(t \in [-h,h]\). Further if \(f_{k-1}\) is continuous on \([-h,h]\), then \(F\bigl(t,f_{k-1}(t)\bigr)\) is continuous as a function of \(t\) on \([-h,h]\) (left as an exercise). Define \[f_k(x) := y_0+ \int_{0}^x F\bigl(t,f_{k-1}(t)\bigr)~dt ,\] and \(f_k\) is continuous on \([-h,h]\) by the fundamental theorem of calculus. To see that \(f_k\) maps \([-h,h]\) to \([y_0-\alpha,y_0+\alpha]\), we compute for \(x \in [-h,h]\) \[\left\lvert {f_k(x) - y_0} \right\rvert = \left\lvert {\int_{0}^x F\bigl(t,f_{k-1}(t)\bigr)~dt } \right\rvert \leq M\left\lvert {x} \right\rvert \leq Mh \leq M \frac{\alpha}{M+L\alpha} \leq \alpha .\] We now define \(f_{k+1}\) and so on, and we have defined a sequence \(\{ f_k \}\) of functions. We need to show that it converges to a function \(f\) that solves the equation [picard:inteq] and therefore [picard:diffeq].

We wish to show that the sequence \(\{ f_k \}\) converges uniformly to some function on \([-h,h]\). First, for \(t \in [-h,h]\) we have the following useful bound \[\left\lvert {F\bigl(t,f_{n}(t)\bigr) - F\bigl(t,f_{k}(t)\bigr)} \right\rvert \leq L \left\lvert {f_n(t)-f_k(t)} \right\rvert \leq L \left\lVert {f_n-f_k} \right\rVert_u ,\] where \(\left\lVert {f_n-f_k} \right\rVert_u\) is the uniform norm, that is the supremum of \(\left\lvert {f_n(t)-f_k(t)} \right\rvert\) for \(t \in [-h,h]\). Now note that \(\left\lvert {x} \right\rvert \leq h \leq \frac{\alpha}{M+L\alpha}\). Therefore \[\begin{split} \left\lvert {f_n(x) - f_k(x)} \right\rvert & = \left\lvert {\int_{0}^x F\bigl(t,f_{n-1}(t)\bigr)~dt - \int_{0}^x F\bigl(t,f_{k-1}(t)\bigr)~dt} \right\rvert \\ & = \left\lvert {\int_{0}^x F\bigl(t,f_{n-1}(t)\bigr)- F\bigl(t,f_{k-1}(t)\bigr)~dt} \right\rvert \\ & \leq L\left\lVert {f_{n-1}-f_{k-1}} \right\rVert_u \left\lvert {x} \right\rvert \\ & \leq \frac{L\alpha}{M+L\alpha} \left\lVert {f_{n-1}-f_{k-1}} \right\rVert_u . \end{split}\] Let \(C := \frac{L\alpha}{M+L\alpha}\) and note that \(C < 1\). Taking supremum on the left-hand side we get \[\left\lVert {f_n-f_k} \right\rVert_u \leq C \left\lVert {f_{n-1}-f_{k-1}} \right\rVert_u .\] Without loss of generality, suppose \(n \geq k\). Then by we can show \[\left\lVert {f_n-f_k} \right\rVert_u \leq C^{k} \left\lVert {f_{n-k}-f_{0}} \right\rVert_u .\] For \(x \in [-h,h]\) we have \[\left\lvert {f_{n-k}(x)-f_{0}(x)} \right\rvert = \left\lvert {f_{n-k}(x)-y_0} \right\rvert \leq \alpha .\] Therefore, \[\left\lVert {f_n-f_k} \right\rVert_u \leq C^{k} \left\lVert {f_{n-k}-f_{0}} \right\rVert_u \leq C^{k} \alpha .\] As \(C < 1\), \(\{f_n\}\) is uniformly Cauchy and by we obtain that \(\{ f_n \}\) converges uniformly on \([-h,h]\) to some function \(f \colon [-h,h] \to {\mathbb{R}}\). The function \(f\) is the uniform limit of continuous functions and therefore continuous. Furthremore since all the \(f_n([-h,h]) \subset [y_0-\alpha,y_0+\alpha]\), then \(f([-h,h]) \subset [y_0-\alpha,y_0+\alpha]\) (why?).

We now need to show that \(f\) solves [picard:inteq]. First, as before we notice \[\left\lvert {F\bigl(t,f_{n}(t)\bigr) - F\bigl(t,f(t)\bigr)} \right\rvert \leq L \left\lvert {f_n(t)-f(t)} \right\rvert \leq L \left\lVert {f_n-f} \right\rVert_u .\] As \(\left\lVert {f_n-f} \right\rVert_u\) converges to 0, then \(F\bigl(t,f_n(t)\bigr)\) converges uniformly to \(F\bigl(t,f(t)\bigr)\) for \(t \in [-h,h]\). Hence, for \(x \in [-h,h]\) the convergence is uniform for \(t \in [0,x]\) (or \([x,0]\) if \(x < 0\)). Therefore, \[\begin{aligned} y_0 + \int_0^{x} F(t,f(t)\bigr)~dt & = y_0 + \int_0^{x} F\bigl(t,\lim_{n\to\infty} f_n(t)\bigr)~dt & & \\ & = y_0 + \int_0^{x} \lim_{n\to\infty} F\bigl(t,f_n(t)\bigr)~dt & & \text{(by continuity of $F$)} \\ & = \lim_{n\to\infty} \left( y_0 + \int_0^{x} F\bigl(t,f_n(t)\bigr)~dt \right) & & \text{(by uniform convergence)} \\ & = \lim_{n\to\infty} f_{n+1}(x) = f(x) . & &\end{aligned}\] We apply the fundamental theorem of calculus to show that \(f\) is differentiable and its derivative is \(F\bigl(x,f(x)\bigr)\). It is obvious that \(f(0) = y_0\).

Finally, what is left to do is to show uniqueness. Suppose \(g \colon [-h,h] \to J \subset {\mathbb{R}}\) is another solution. As before we use the fact that \(\left\lvert {F\bigl(t,f(t)\bigr) - F\bigl(t,g(t)\bigr)} \right\rvert \leq L \left\lVert {f-g} \right\rVert_u\). Then \[\begin{split} \left\lvert {f(x)-g(x)} \right\rvert & = \left\lvert { y_0 + \int_0^{x} F\bigl(t,f(t)\bigr)~dt - \left( y_0 + \int_0^{x} F\bigl(t,g(t)\bigr)~dt \right) } \right\rvert \\ & = \left\lvert { \int_0^{x} F\bigl(t,f(t)\bigr) - F\bigl(t,g(t)\bigr)~dt } \right\rvert \\ & \leq L\left\lVert {f-g} \right\rVert_u\left\lvert {x} \right\rvert \leq Lh\left\lVert {f-g} \right\rVert_u \leq \frac{L\alpha}{M+L\alpha}\left\lVert {f-g} \right\rVert_u . \end{split}\] As before, \(C = \frac{L\alpha}{M+L\alpha} < 1\). By taking supremum over \(x \in [-h,h]\) on the left hand side we obtain \[\left\lVert {f-g} \right\rVert_u \leq C \left\lVert {f-g} \right\rVert_u .\] This is only possible if \(\left\lVert {f-g} \right\rVert_u = 0\). Therefore, \(f=g\), and the solution is unique.

#### Examples

Let us look at some examples. The proof of the theorem gives us an explicit way to find an \(h\) that works. It does not, however, give us the best \(h\). It is often possible to find a much larger \(h\) for which the conclusion of the theorem holds.

The proof also gives us the Picard iterates as approximations to the solution. So the proof actually tells us how to obtain the solution, not just that the solution exists.

Consider \[f'(x) = f(x), \qquad f(0) = 1 .\] That is, we let \(F(x,y) = y\), and we are looking for a function \(f\) such that \(f'(x) = f(x)\). We pick any \(I\) that contains 0 in the interior. We pick an arbitrary \(J\) that contains 1 in its interior. We can use \(L = 1\). The theorem guarantees an \(h > 0\) such that there exists a unique solution \(f \colon [-h,h] \to {\mathbb{R}}\). This solution is usually denoted by \[e^x := f(x) .\] We leave it to the reader to verify that by picking \(I\) and \(J\) large enough the proof of the theorem guarantees that we are able to pick \(\alpha\) such that we get any \(h\) we want as long as \(h < \nicefrac{1}{2}\). We omit the calculation.

Of course, we know this function exists as a function for all \(x\), so an arbitrary \(h\) ought to work. By same reasoning as above, no matter what \(x_0\) and \(y_0\) are, the proof guarantees an arbitrary \(h\) as long as \(h < \nicefrac{1}{2}\). Fix such an \(h\). We get a unique function defined on \([x_0-h,x_0+h]\). After defining the function on \([-h,h]\) we find a solution on the interval \([0,2h]\) and notice that the two functions must coincide on \([0,h]\) by uniqueness. We thus iteratively construct the exponential for all \(x \in {\mathbb{R}}\). Therefore Picard’s theorem could be used to prove the existence and uniqueness of the exponential.

Let us compute the Picard iterates. We start with the constant function \(f_0(x) := 1\). Then \[\begin{aligned} f_1(x) & = 1 + \int_0^x f_0(s)~ds = 1+x, \\ f_2(x) & = 1 + \int_0^x f_1(s)~ds = 1 + \int_0^x (1+s)~ds = 1 + x + \frac{x^2}{2}, \\ f_3(x) & = 1 + \int_0^x f_2(s)~ds = 1 + \int_0^x \left(1+ s + \frac{s^2}{2} \right)~ds = 1 + x + \frac{x^2}{2} + \frac{x^3}{6} .\end{aligned}\] We recognize the beginning of the Taylor series for the exponential.

Suppose we have the equation \[f'(x) = {\bigl(f(x)\bigr)}^2 \qquad \text{and} \qquad f(0)=1.\] From elementary differential equations we know \[f(x) = \frac{1}{1-x}\] is the solution. The solution is only defined on \((-\infty,1)\). That is, we are able to use \(h < 1\), but never a larger \(h\). The function that takes \(y\) to \(y^2\) is not Lipschitz as a function on all of \({\mathbb{R}}\). As we approach \(x=1\) from the left, the solution becomes larger and larger. The derivative of the solution grows as \(y^2\), and therefore the \(L\) required will have to be larger and larger as \(y_0\) grows. Thus if we apply the theorem with \(x_0\) close to 1 and \(y_0 = \frac{1}{1-x_0}\) we find that the \(h\) that the proof guarantees will be smaller and smaller as \(x_0\) approaches 1.

By picking \(\alpha\) correctly, the proof of the theorem guarantees \(h=1-\nicefrac{\sqrt{3}}{2} \approx 0.134\) (we omit the calculation) for \(x_0=0\) and \(y_0=1\), even though we saw above that any \(h < 1\) should work.

Consider the equation \[f'(x) = 2 \sqrt{\left\lvert {f(x)} \right\rvert}, \qquad f(0) = 0 .\] The function \(F(x,y) = 2 \sqrt{\left\lvert {y} \right\rvert}\) is continuous, but not Lipschitz in \(y\) (why?). The equation does not satisfy the hypotheses of the theorem. The function \[f(x) = \begin{cases} x^2 & \text{ if $x \geq 0$,}\\ -x^2 & \text{ if $x < 0$,} \end{cases}\] is a solution, but \(f(x) = 0\) is also a solution. A solution exists, but is not unique.

Consider \(y' = \varphi(x)\) where \(\varphi(x) := 0\) if \(x \in {\mathbb{Q}}\) and \(\varphi(x):=1\) if \(x \not\in {\mathbb{Q}}\). The equation has no solution regardless of the initial conditions. A solution would have derivative \(\varphi\), but \(\varphi\) does not have the intermediate value property at any point (why?). No solution exists by . Therefore to obtain existence of a solution, some continuity hypothesis on \(F\) is necessary.

## Exercises

Let \(I, J \subset {\mathbb{R}}\) be intervals. Let \(F \colon I \times J \to {\mathbb{R}}\) be a continuous function of two variables and suppose \(f \colon I \to J\) be a continuous function. Show that \(F\bigl(x,f(x)\bigr)\) is a continuous function on \(I\).

Let \(I, J \subset {\mathbb{R}}\) be closed bounded intervals. Show that if \(F \colon I \times J \to {\mathbb{R}}\) is continuous, then \(F\) is bounded.

We proved Picard’s theorem under the assumption that \(x_0 = 0\). Prove the full statement of Picard’s theorem for an arbitrary \(x_0\).

Let \(f'(x)=x f(x)\) be our equation. Start with the initial condition \(f(0)=2\) and find the Picard iterates \(f_0,f_1,f_2,f_3,f_4\).

Suppose \(F \colon I \times J \to {\mathbb{R}}\) is a function that is continuous in the first variable, that is, for any fixed \(y\) the function that takes \(x\) to \(F(x,y)\) is continuous. Further, suppose \(F\) is Lipschitz in the second variable, that is, there exists a number \(L\) such that \[\left\lvert {F(x,y) - F(x,z)} \right\rvert \leq L \left\lvert {y-z} \right\rvert \ \ \ \text{ for all $y,z \in J$, $x \in I$} .\] Show that \(F\) is continuous as a function of two variables. Therefore, the hypotheses in the theorem could be made even weaker.

A common type of equation one encounters are *linear first order differential equations*, that is equations of the form \[y' + p(x) y = q(x) , \qquad y(x_0) = y_0 .\] Prove Picard’s theorem for linear equations. Suppose \(I\) is an interval, \(x_0 \in I\), and \(p \colon I \to {\mathbb{R}}\) and \(q \colon I \to {\mathbb{R}}\) are continuous. Show that there exists a unique differentiable \(f \colon I \to {\mathbb{R}}\), such that \(y = f(x)\) satisfies the equation and the initial condition. Hint: Assume existence of the exponential function and use the integrating factor formula for existence of \(f\) (prove that it works): \[f(x) := e^{-\int_{x_0}^x p(s)\, ds} \left( \int_{x_0}^x e^{\int_{x_0}^t p(s)\, ds} q(t) ~dt + y_0 \right).\]

## Metric Spaces

### Metric spaces

*Note: 1.5 lectures*

As mentioned in the introduction, the main idea in analysis is to take limits. In we learned to take limits of sequences of real numbers. And in we learned to take limits of functions as a real number approached some other real number.

We want to take limits in more complicated contexts. For example, we want to have sequences of points in 3-dimensional space. We wish to define continuous functions of several variables. We even want to define functions on spaces that are a little harder to describe, such as the surface of the earth. We still want to talk about limits there.

Finally, we have seen the limit of a sequence of functions in . We wish to unify all these notions so that we do not have to reprove theorems over and over again in each context. The concept of a metric space is an elementary yet powerful tool in analysis. And while it is not sufficient to describe every type of limit we find in modern analysis, it gets us very far indeed.

Let \(X\) be a set, and let \(d \colon X \times X \to {\mathbb{R}}\) be a function such that

- [metric:pos] \(d(x,y) \geq 0\) for all \(x, y\) in \(X\),
- [metric:zero] \(d(x,y) = 0\) if and only if \(x = y\),
- [metric:com] \(d(x,y) = d(y,x)\),
- [metric:triang] \(d(x,z) \leq d(x,y)+ d(y,z)\) (
*triangle inequality*).

Then the pair \((X,d)\) is called a *metric space*. The function \(d\) is called the *metric* or sometimes the *distance function*. Sometimes we just say \(X\) is a metric space if the metric is clear from context.

The geometric idea is that \(d\) is the distance between two points. Items [metric:pos]–[metric:com] have obvious geometric interpretation: distance is always nonnegative, the only point that is distance 0 away from \(x\) is \(x\) itself, and finally that the distance from \(x\) to \(y\) is the same as the distance from \(y\) to \(x\). The triangle inequality [metric:triang] has the interpretation given in .

For the purposes of drawing, it is convenient to draw figures and diagrams in the plane and have the metric be the standard distance. However, that is only one particular metric space. Just because a certain fact seems to be clear from drawing a picture does not mean it is true. You might be getting sidetracked by intuition from euclidean geometry, whereas the concept of a metric space is a lot more general.

Let us give some examples of metric spaces.

The set of real numbers \({\mathbb{R}}\) is a metric space with the metric \[d(x,y) := \left\lvert {x-y} \right\rvert .\] Items [metric:pos]–[metric:com] of the definition are easy to verify. The triangle inequality [metric:triang] follows immediately from the standard triangle inequality for real numbers: \[d(x,z) = \left\lvert {x-z} \right\rvert = \left\lvert {x-y+y-z} \right\rvert \leq \left\lvert {x-y} \right\rvert+\left\lvert {y-z} \right\rvert = d(x,y)+ d(y,z) .\] This metric is the *standard metric on \({\mathbb{R}}\)*. If we talk about \({\mathbb{R}}\) as a metric space without mentioning a specific metric, we mean this particular metric.

We can also put a different metric on the set of real numbers. For example, take the set of real numbers \({\mathbb{R}}\) together with the metric \[d(x,y) := \frac{\left\lvert {x-y} \right\rvert}{\left\lvert {x-y} \right\rvert+1} .\] Items [metric:pos]–[metric:com] are again easy to verify. The triangle inequality [metric:triang] is a little bit more difficult. Note that \(d(x,y) = \varphi(\left\lvert {x-y} \right\rvert)\) where \(\varphi(t) = \frac{t}{t+1}\) and \(\varphi\) is an increasing function (positive derivative). Hence \[\begin{split} d(x,z) & = \varphi(\left\lvert {x-z} \right\rvert) = \varphi(\left\lvert {x-y+y-z} \right\rvert) \leq \varphi(\left\lvert {x-y} \right\rvert+\left\lvert {y-z} \right\rvert) \\ & = \frac{\left\lvert {x-y} \right\rvert+\left\lvert {y-z} \right\rvert}{\left\lvert {x-y} \right\rvert+\left\lvert {y-z} \right\rvert+1} = \frac{\left\lvert {x-y} \right\rvert}{\left\lvert {x-y} \right\rvert+\left\lvert {y-z} \right\rvert+1} + \frac{\left\lvert {y-z} \right\rvert}{\left\lvert {x-y} \right\rvert+\left\lvert {y-z} \right\rvert+1} \\ & \leq \frac{\left\lvert {x-y} \right\rvert}{\left\lvert {x-y} \right\rvert+1} + \frac{\left\lvert {y-z} \right\rvert}{\left\lvert {y-z} \right\rvert+1} = d(x,y)+ d(y,z) . \end{split}\] Here we have an example of a nonstandard metric on \({\mathbb{R}}\). With this metric we see for example that \(d(x,y) < 1\) for all \(x,y \in {\mathbb{R}}\). That is, any two points are less than 1 unit apart.

An important metric space is the \(n\)-dimensional *euclidean space* \({\mathbb{R}}^n = {\mathbb{R}} \times {\mathbb{R}}\times \cdots \times {\mathbb{R}}\). We use the following notation for points: \(x =(x_1,x_2,\ldots,x_n) \in {\mathbb{R}}^n\). We also simply write \(0 \in {\mathbb{R}}^n\) to mean the vector \((0,0,\ldots,0)\). Before making \({\mathbb{R}}^n\) a metric space, let us prove an important inequality, the so-called Cauchy-Schwarz inequality.

Take \(x =(x_1,x_2,\ldots,x_n) \in {\mathbb{R}}^n\) and \(y =(y_1,y_2,\ldots,y_n) \in {\mathbb{R}}^n\). Then \[{\biggl( \sum_{j=1}^n x_j y_j \biggr)}^2 \leq \biggl(\sum_{j=1}^n x_j^2 \biggr) \biggl(\sum_{j=1}^n y_j^2 \biggr) .\]

Any square of a real number is nonnegative. Hence any sum of squares is nonnegative: \[\begin{split} 0 & \leq \sum_{j=1}^n \sum_{k=1}^n {(x_j y_k - x_k y_j)}^2 \\ & = \sum_{j=1}^n \sum_{k=1}^n \bigl( x_j^2 y_k^2 + x_k^2 y_j^2 - 2 x_j x_k y_j y_k \bigr) \\ & = \biggl( \sum_{j=1}^n x_j^2 \biggr) \biggl( \sum_{k=1}^n y_k^2 \biggr) + \biggl( \sum_{j=1}^n y_j^2 \biggr) \biggl( \sum_{k=1}^n x_k^2 \biggr) - 2 \biggl( \sum_{j=1}^n x_j y_j \biggr) \biggl( \sum_{k=1}^n x_k y_k \biggr) \end{split}\] We relabel and divide by 2 to obtain \[0 \leq \biggl( \sum_{j=1}^n x_j^2 \biggr) \biggl( \sum_{j=1}^n y_j^2 \biggr) - {\biggl( \sum_{j=1}^n x_j y_j \biggr)}^2 ,\] which is precisely what we wanted.

Let us construct the standard metric for \({\mathbb{R}}^n\). Define \[d(x,y) := \sqrt{ {(x_1-y_1)}^2 + {(x_2-y_2)}^2 + \cdots + {(x_n-y_n)}^2 } = \sqrt{ \sum_{j=1}^n {(x_j-y_j)}^2 } .\] For \(n=1\), the real line, this metric agrees with what we did above. Again, the only tricky part of the definition to check is the triangle inequality. It is less messy to work with the square of the metric. In the following, note the use of the Cauchy-Schwarz inequality. \[\begin{split} {\bigl(d(x,z)\bigr)}^2 & = \sum_{j=1}^n {(x_j-z_j)}^2 \\ & = \sum_{j=1}^n {(x_j-y_j+y_j-z_j)}^2 \\ & = \sum_{j=1}^n \Bigl( {(x_j-y_j)}^2+{(y_j-z_j)}^2 + 2(x_j-y_j)(y_j-z_j) \Bigr) \\ & = \sum_{j=1}^n {(x_j-y_j)}^2 + \sum_{j=1}^n {(y_j-z_j)}^2 + \sum_{j=1}^n 2(x_j-y_j)(y_j-z_j) \\ & \leq \sum_{j=1}^n {(x_j-y_j)}^2 + \sum_{j=1}^n {(y_j-z_j)}^2 + 2 \sqrt{ \sum_{j=1}^n {(x_j-y_j)}^2 \sum_{j=1}^n {(y_j-z_j)}^2 } \\ & = {\left( \sqrt{ \sum_{j=1}^n {(x_j-y_j)}^2 } + \sqrt{ \sum_{j=1}^n {(y_j-z_j)}^2 } \right)}^2 = {\bigl( d(x,y) + d(y,z) \bigr)}^2 . \end{split}\] Taking the square root of both sides we obtain the correct inequality, because the square root is an increasing function.

An example to keep in mind is the so-called *discrete metric*. Let \(X\) be any set and define \[d(x,y) := \begin{cases} 1 & \text{if $x \not= y$}, \\ 0 & \text{if $x = y$}. \end{cases}\] That is, all points are equally distant from each other. When \(X\) is a finite set, we can draw a diagram, see for example . Things become subtle when \(X\) is an infinite set such as the real numbers.

While this particular example seldom comes up in practice, it gives a useful “smell test.” If you make a statement about metric spaces, try it with the discrete metric. To show that \((X,d)\) is indeed a metric space is left as an exercise.

[example:msC01] Let \(C([a,b],{\mathbb{R}})\) be the set of continuous real-valued functions on the interval \([a,b]\). Define the metric on \(C([a,b],{\mathbb{R}})\) as \[d(f,g) := \sup_{x \in [a,b]} \left\lvert {f(x)-g(x)} \right\rvert .\] Let us check the properties. First, \(d(f,g)\) is finite as \(\left\lvert {f(x)-g(x)} \right\rvert\) is a continuous function on a closed bounded interval \([a,b]\), and so is bounded. It is clear that \(d(f,g) \geq 0\), it is the supremum of nonnegative numbers. If \(f = g\) then \(\left\lvert {f(x)-g(x)} \right\rvert = 0\) for all \(x\) and hence \(d(f,g) = 0\). Conversely if \(d(f,g) = 0\), then for any \(x\) we have \(\left\lvert {f(x)-g(x)} \right\rvert \leq d(f,g) = 0\) and hence \(f(x) = g(x)\) for all \(x\) and \(f=g\). That \(d(f,g) = d(g,f)\) is equally trivial. To show the triangle inequality we use the standard triangle inequality. \[\begin{split} d(f,g) & = \sup_{x \in [a,b]} \left\lvert {f(x)-g(x)} \right\rvert = \sup_{x \in [a,b]} \left\lvert {f(x)-h(x)+h(x)-g(x)} \right\rvert \\ & \leq \sup_{x \in [a,b]} ( \left\lvert {f(x)-h(x)} \right\rvert+\left\lvert {h(x)-g(x)} \right\rvert ) \\ & \leq \sup_{x \in [a,b]} \left\lvert {f(x)-h(x)} \right\rvert+ \sup_{x \in [a,b]} \left\lvert {h(x)-g(x)} \right\rvert = d(f,h) + d(h,g) . \end{split}\] When treating \(C([a,b],{\mathbb{R}})\) as a metric space without mentioning a metric, we mean this particular metric. Notice that \(d(f,g) = \left\lVert {f-g} \right\rVert_u\), the uniform norm of .

This example may seem esoteric at first, but it turns out that working with spaces such as \(C([a,b],{\mathbb{R}})\) is really the meat of a large part of modern analysis. Treating sets of functions as metric spaces allows us to abstract away a lot of the grubby detail and prove powerful results such as Picard’s theorem with less work.

Oftentimes it is useful to consider a subset of a larger metric space as a metric space itself. We obtain the following proposition, which has a trivial proof.

Let \((X,d)\) be a metric space and \(Y \subset X\), then the restriction \(d|_{Y \times Y}\) is a metric on \(Y\).

If \((X,d)\) is a metric space, \(Y \subset X\), and \(d' := d|_{Y \times Y}\), then \((Y,d')\) is said to be a *subspace* of \((X,d)\).

It is common to simply write \(d\) for the metric on \(Y\), as it is the restriction of the metric on \(X\). Sometimes we say \(d'\) is the *subspace metric* and \(Y\) has the *subspace topology*.

A subset of the real numbers is bounded whenever all its elements are at most some fixed distance from 0. We also define bounded sets in a metric space. When dealing with an arbitrary metric space there may not be some natural fixed point 0. For the purposes of boundedness it does not matter.

Let \((X,d)\) be a metric space. A subset \(S \subset X\) is said to be *bounded* if there exists a \(p \in X\) and a \(B \in {\mathbb{R}}\) such that \[d(p,x) \leq B \quad \text{for all $x \in S$}.\] We say \((X,d)\) is bounded if \(X\) itself is a bounded subset.

For example, the set of real numbers with the standard metric is not a bounded metric space. It is not hard to see that a subset of the real numbers is bounded in the sense of if and only if it is bounded as a subset of the metric space of real numbers with the standard metric.

On the other hand, if we take the real numbers with the discrete metric, then we obtain a bounded metric space. In fact, any set with the discrete metric is bounded.

## Exercises

Show that for any set \(X\), the discrete metric (\(d(x,y) = 1\) if \(x\not=y\) and \(d(x,x) = 0\)) does give a metric space \((X,d)\).

Let \(X := \{ 0 \}\) be a set. Can you make it into a metric space?

Let \(X := \{ a, b \}\) be a set. Can you make it into two distinct metric spaces? (define two distinct metrics on it)

Let the set \(X := \{ A, B, C \}\) represent 3 buildings on campus. Suppose we wish our distance to be the time it takes to walk from one building to the other. It takes 5 minutes either way between buildings \(A\) and \(B\). However, building \(C\) is on a hill and it takes 10 minutes from \(A\) and 15 minutes from \(B\) to get to \(C\). On the other hand it takes 5 minutes to go from \(C\) to \(A\) and 7 minutes to go from \(C\) to \(B\), as we are going downhill. Do these distances define a metric? If so, prove it, if not, say why not.

Suppose \((X,d)\) is a metric space and \(\varphi \colon [0,\infty) \to {\mathbb{R}}\) is a function such that \(\varphi(t) \geq 0\) for all \(t\) and \(\varphi(t) = 0\) if and only if \(t=0\). Also suppose \(\varphi\) is *subadditive*, that is, \(\varphi(s+t) \leq \varphi(s)+\varphi(t)\). Show that with \(d'(x,y) := \varphi\bigl(d(x,y)\bigr)\), we obtain a new metric space \((X,d')\).

[exercise:mscross] Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces.

a) Show that \((X \times Y,d)\) with \(d\bigl( (x_1,y_1), (x_2,y_2) \bigr) := d_X(x_1,x_2) + d_Y(y_1,y_2)\) is a metric space.

b) Show that \((X \times Y,d)\) with \(d\bigl( (x_1,y_1), (x_2,y_2) \bigr) := \max \{ d_X(x_1,x_2) , d_Y(y_1,y_2) \}\) is a metric space.

Let \(X\) be the set of continuous functions on \([0,1]\). Let \(\varphi \colon [0,1] \to (0,\infty)\) be continuous. Define \[d(f,g) := \int_0^1 \left\lvert {f(x)-g(x)} \right\rvert\varphi(x)~dx .\] Show that \((X,d)\) is a metric space.

[exercise:mshausdorffpseudo] Let \((X,d)\) be a metric space. For nonempty bounded subsets \(A\) and \(B\) let \[d(x,B) := \inf \{ d(x,b) : b \in B \} \qquad \text{and} \qquad d(A,B) := \sup \{ d(a,B) : a \in A \} .\] Now define the *Hausdorff metric* as \[d_H(A,B) := \max \{ d(A,B) , d(B,A) \} .\] Note: \(d_H\) can be defined for arbitrary nonempty subsets if we allow the extended reals.

a) Let \(Y \subset {\mathcal{P}}(X)\) be the set of bounded nonempty subsets. Prove that \((Y,d_H)\) is a so-called *pseudometric space*: \(d_H\) satisfies the metric properties [metric:pos], [metric:com], [metric:triang], and further \(d_H(A,A) = 0\) for all \(A \in Y\).

b) Show by example that \(d\) itself is not symmetric, that is \(d(A,B) \not= d(B,A)\).

c) Find a metric space \(X\) and two different nonempty bounded subsets \(A\) and \(B\) such that \(d_H(A,B) = 0\).

### Open and closed sets

*Note: 2 lectures*

#### Topology

It is useful to define a so-called *topology*. That is we define closed and open sets in a metric space. Before doing so, let us define two special sets.

Let \((X,d)\) be a metric space, \(x \in X\) and \(\delta > 0\). Then define the *open ball* or simply *ball* of radius \(\delta\) around \(x\) as \[B(x,\delta) := \{ y \in X : d(x,y) < \delta \} .\] Similarly we define the *closed ball* as \[C(x,\delta) := \{ y \in X : d(x,y) \leq \delta \} .\]

When we are dealing with different metric spaces, it is sometimes convenient to emphasize which metric space the ball is in. We do this by writing \(B_X(x,\delta) := B(x,\delta)\) or \(C_X(x,\delta) := C(x,\delta)\).

Take the metric space \({\mathbb{R}}\) with the standard metric. For \(x \in {\mathbb{R}}\), and \(\delta > 0\) we get \[B(x,\delta) = (x-\delta,x+\delta) \qquad \text{and} \qquad C(x,\delta) = [x-\delta,x+\delta] .\]

Be careful when working on a subspace. Suppose we take the metric space \([0,1]\) as a subspace of \({\mathbb{R}}\). Then in \([0,1]\) we get \[B(0,\nicefrac{1}{2}) = B_{[0,1]}(0,\nicefrac{1}{2}) = [0,\nicefrac{1}{2}) .\] This is different from \(B_

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[39]/div/div[1]/p[5]/span, line 1, column 1 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Let \((X,d)\) be a metric space. A set \(V \subset X\) is *open* if for every \(x \in V\), there exists a \(\delta > 0\) such that \(B(x,\delta) \subset V\). See . A set \(E \subset X\) is *closed* if the complement \(E^c = X \setminus E\) is open. When the ambient space \(X\) is not clear from context we say \(V\) is open in \(X\) and \(E\) is closed in \(X\).

If \(x \in V\) and \(V\) is open, then we say \(V\) is an *open neighborhood* of \(x\) (or sometimes just *neighborhood*).

Intuitively, an open set is a set that does not include its “boundary,” wherever we are at in the set, we are allowed to “wiggle” a little bit and stay in the set. Note that not every set is either open or closed, in fact generally most subsets are neither.

The set \([0,1) \subset {\mathbb{R}}\) is neither open nor closed. First, every ball in \({\mathbb{R}}\) around \(0\), \((-\delta,\delta)\), contains negative numbers and hence is not contained in \([0,1)\) and so \([0,1)\) is not open. Second, every ball in \({\mathbb{R}}\) around \(1\), \((1-\delta,1+\delta)\) contains numbers strictly less than 1 and greater than 0 (e.g. \(1-\nicefrac{\delta}{2}\) as long as \(\delta < 2\)). Thus \({\mathbb{R}}\setminus [0,1)\) is not open, and so \([0,1)\) is not closed.

[prop:topology:open] Let \((X,d)\) be a metric space.

- [topology:openi] \(\emptyset\) and \(X\) are open in \(X\).
- [topology:openii] If \(V_1, V_2, \ldots, V_k\) are open then \[\bigcap_{j=1}^k V_j\] is also open. That is, finite intersection of open sets is open.
- [topology:openiii] If \(\{ V_\lambda \}_{\lambda \in I}\) is an arbitrary collection of open sets, then \[\bigcup_{\lambda \in I} V_\lambda\] is also open. That is, union of open sets is open.

Note that the index set in [topology:openiii] is arbitrarily large. By \(\bigcup_{\lambda \in I} V_\lambda\) we simply mean the set of all \(x\) such that \(x \in V_\lambda\) for at least one \(\lambda \in I\).

The sets \(X\) and \(\emptyset\) are obviously open in \(X\).

Let us prove [topology:openii]. If \(x \in \bigcap_{j=1}^k V_j\), then \(x \in V_j\) for all \(j\). As \(V_j\) are all open, for every \(j\) there exists a \(\delta_j > 0\) such that \(B(x,\delta_j) \subset V_j\). Take \(\delta := \min \{ \delta_1,\delta_2,\ldots,\delta_k \}\) and notice \(\delta > 0\). We have \(B(x,\delta) \subset B(x,\delta_j) \subset V_j\) for every \(j\) and so \(B(x,\delta) \subset \bigcap_{j=1}^k V_j\). Consequently the intersection is open.

Let us prove [topology:openiii]. If \(x \in \bigcup_{\lambda \in I} V_\lambda\), then \(x \in V_\lambda\) for some \(\lambda \in I\). As \(V_\lambda\) is open, there exists a \(\delta > 0\) such that \(B(x,\delta) \subset V_\lambda\). But then \(B(x,\delta) \subset \bigcup_{\lambda \in I} V_\lambda\) and so the union is open.

The main thing to notice is the difference between items [topology:openii] and [topology:openiii]. Item [topology:openii] is not true for an arbitrary intersection, for example \(\bigcap_{n=1}^\infty (-\nicefrac{1}{n},\nicefrac{1}{n}) = \{ 0 \}\), which is not open.

The proof of the following analogous proposition for closed sets is left as an exercise.

[prop:topology:closed] Let \((X,d)\) be a metric space.

- [topology:closedi] \(\emptyset\) and \(X\) are closed in \(X\).
- [topology:closedii] If \(\{ E_\lambda \}_{\lambda \in I}\) is an arbitrary collection of closed sets, then \[\bigcap_{\lambda \in I} E_\lambda\] is also closed. That is, intersection of closed sets is closed.
- [topology:closediii] If \(E_1, E_2, \ldots, E_k\) are closed then \[\bigcup_{j=1}^k E_j\] is also closed. That is, finite union of closed sets is closed.

We have not yet shown that the open ball is open and the closed ball is closed. Let us show this fact now to justify the terminology.

[prop:topology:ballsopenclosed] Let \((X,d)\) be a metric space, \(x \in X\), and \(\delta > 0\). Then \(B(x,\delta)\) is open and \(C(x,\delta)\) is closed.

Let \(y \in B(x,\delta)\). Let \(\alpha := \delta-d(x,y)\). Of course \(\alpha > 0\). Now let \(z \in B(y,\alpha)\). Then \[d(x,z) \leq d(x,y) + d(y,z) < d(x,y) + \alpha = d(x,y) + \delta-d(x,y) = \delta .\] Therefore \(z \in B(x,\delta)\) for every \(z \in B(y,\alpha)\). So \(B(y,\alpha) \subset B(x,\delta)\) and \(B(x,\delta)\) is open.

The proof that \(C(x,\delta)\) is closed is left as an exercise.

Again be careful about what is the ambient metric space. As \([0,\nicefrac{1}{2})\) is an open ball in \([0,1]\), this means that \([0,\nicefrac{1}{2})\) is an open set in \([0,1]\). On the other hand \([0,\nicefrac{1}{2})\) is neither open nor closed in \({\mathbb{R}}\).

A useful way to think about an open set is as a union of open balls. If \(U\) is open, then for each \(x \in U\), there is a \(\delta_x > 0\) (depending on \(x\)) such that \(B(x,\delta_x) \subset U\). Then \(U = \bigcup_{x\in U} B(x,\delta_x)\).

The proof of the following proposition is left as an exercise. Note that there are many other open and closed sets in \({\mathbb{R}}\).

[prop:topology:intervals:openclosed] Let \(a < b\) be two real numbers. Then \((a,b)\), \((a,\infty)\), and \((-\infty,b)\) are open in \({\mathbb{R}}\). Also \([a,b]\), \([a,\infty)\), and \((-\infty,b]\) are closed in \({\mathbb{R}}\).

#### Connected sets

A nonempty metric space \((X,d)\) is *connected* if the only subsets of \(X\) that are both open and closed are \(\emptyset\) and \(X\) itself. If \((X,d)\) is not connected we say it is *disconnected*.

When we apply the term *connected* to a nonempty subset \(A \subset X\), we simply mean that \(A\) with the subspace topology is connected.

In other words, a nonempty \(X\) is connected if whenever we write \(X = X_1 \cup X_2\) where \(X_1 \cap X_2 = \emptyset\) and \(X_1\) and \(X_2\) are open, then either \(X_1 = \emptyset\) or \(X_2 = \emptyset\). So to show \(X\) is disconnected, we need to find nonempty disjoint open sets \(X_1\) and \(X_2\) whose union is \(X\). For subsets, we state this idea as a proposition.

Let \((X,d)\) be a metric space. A nonempty set \(S \subset X\) is not connected if and only if there exist open sets \(U_1\) and \(U_2\) in \(X\), such that \(U_1 \cap U_2 \cap S = \emptyset\), \(U_1 \cap S \not= \emptyset\), \(U_2 \cap S \not= \emptyset\), and \[S = \bigl( U_1 \cap S \bigr) \cup \bigl( U_2 \cap S \bigr) .\]

If \(U_j\) is open in \(X\), then \(U_j \cap S\) is open in \(S\) in the subspace topology (with subspace metric). To see this, note that if \(B_X(x,\delta) \subset U_j\), then as \(B_S(x,\delta) = S \cap B_X(x,\delta)\), we have \(B_S(x,\delta) \subset U_j \cap S\). So if \(U_1\) and \(U_2\) as above exist, then \(X\) is disconnected based on the discussion above.

The proof of the other direction follows by using to find \(U_1\) and \(U_2\) from two open disjoint subsets of \(S\).

Let \(S \subset {\mathbb{R}}\) be such that \(x < z < y\) with \(x,y \in S\) and \(z \notin S\). Claim: \(S\) is not connected. Proof: Notice \[\bigl( (-\infty,z) \cap S \bigr) \cup \bigl( (z,\infty) \cap S \bigr) = S .\]

A nonempty set \(S \subset {\mathbb{R}}\) is connected if and only if it is an interval or a single point.

Suppose \(S\) is connected. If \(S\) is a single point then we are done. So suppose \(x < y\) and \(x,y \in S\). If \(z\) is such that \(x < z < y\), then \((-\infty,z) \cap S\) is nonempty and \((z,\infty) \cap S\) is nonempty. The two sets are disjoint. As \(S\) is connected, we must have they their union is not \(S\), so \(z \in S\).

Suppose \(S\) is bounded, connected, but not a single point. Let \(\alpha := \inf \, S\) and \(\beta := \sup \, S\) and note that \(\alpha < \beta\). Suppose \(\alpha < z < \beta\). As \(\alpha\) is the infimum, then there is an \(x \in S\) such that \(\alpha \leq x < z\). Similarly there is a \(y \in S\) such that \(\beta \geq y > z\). We have shown above that \(z \in S\), so \((\alpha,\beta) \subset S\). If \(w < \alpha\), then \(w \notin S\) as \(\alpha\) was the infimum, similarly if \(w > \beta\) then \(w \notin S\). Therefore the only possibilities for \(S\) are \((\alpha,\beta)\), \([\alpha,\beta)\), \((\alpha,\beta]\), \([\alpha,\beta]\).

The proof that an unbounded connected \(S\) is an interval is left as an exercise.

On the other hand suppose \(S\) is an interval. Suppose \(U_1\) and \(U_2\) are open subsets of \({\mathbb{R}}\), \(U_1 \cap S\) and \(U_2 \cap S\) are nonempty, and \(S = \bigl( U_1 \cap S \bigr) \cup \bigl( U_2 \cap S \bigr)\). We will show that \(U_1 \cap S\) and \(U_2 \cap S\) contain a common point, so they are not disjoint, and hence \(S\) must be connected. Suppose there is \(x \in U_1 \cap S\) and \(y \in U_2 \cap S\). Without loss of generality, assume \(x < y\). As \(S\) is an interval \([x,y] \subset S\). Let \(z := \inf (U_2 \cap [x,y])\). If \(z = x\), then \(z \in U_1\). If \(z > x\), then for any \(\delta > 0\) the ball \(B(z,\delta) = (z-\delta,z+\delta)\) contains points of \(S\) that are not in \(U_2\), and so \(z \notin U_2\) as \(U_2\) is open. Therefore, \(z \in U_1\). As \(U_1\) is open, \(B(z,\delta) \subset U_1\) for a small enough \(\delta > 0\). As \(z\) is the infimum of \(U_2 \cap [x,y]\), there must exist some \(w \in U_2 \cap [x,y]\) such that \(w \in [z,z+\delta) \subset B(z,\delta) \subset U_1\). Therefore \(w \in U_1 \cap U_2 \cap [x,y]\). So \(U_1 \cap S\) and \(U_2 \cap S\) are not disjoint and hence \(S\) is connected.

In many cases a ball \(B(x,\delta)\) is connected. But this is not necessarily true in every metric space. For a simplest example, take a two point space \(\{ a, b\}\) with the discrete metric. Then \(B(a,2) = \{ a , b \}\), which is not connected as \(B(a,1) = \{ a \}\) and \(B(b,1) = \{ b \}\) are open and disjoint.

#### Closure and boundary

Sometimes we wish to take a set and throw in everything that we can approach from the set. This concept is called the closure.

Let \((X,d)\) be a metric space and \(A \subset X\). Then the *closure* of \(A\) is the set \[\overline{A} := \bigcap \{ E \subset X : \text{$E$ is closed and $A \subset E$} \} .\] That is, \(\overline{A}\) is the intersection of all closed sets that contain \(A\).

Let \((X,d)\) be a metric space and \(A \subset X\). The closure \(\overline{A}\) is closed. Furthermore if \(A\) is closed then \(\overline{A} = A\).

First, the closure is the intersection of closed sets, so it is closed. Second, if \(A\) is closed, then take \(E = A\), hence the intersection of all closed sets \(E\) containing \(A\) must be equal to \(A\).

The closure of \((0,1)\) in \({\mathbb{R}}\) is \([0,1]\). Proof: Simply notice that if \(E\) is closed and contains \((0,1)\), then \(E\) must contain \(0\) and \(1\) (why?). Thus \([0,1] \subset E\). But \([0,1]\) is also closed. Therefore the closure \(\overline{(0,1)} = [0,1]\).

Be careful to notice what ambient metric space you are working with. If \(X = (0,\infty)\), then the closure of \((0,1)\) in \((0,\infty)\) is \((0,1]\). Proof: Similarly as above \((0,1]\) is closed in \((0,\infty)\) (why?). Any closed set \(E\) that contains \((0,1)\) must contain 1 (why?). Therefore \((0,1] \subset E\), and hence \(\overline{(0,1)} = (0,1]\) when working in \((0,\infty)\).

Let us justify the statement that the closure is everything that we can “approach” from the set.

[prop:msclosureappr] Let \((X,d)\) be a metric space and \(A \subset X\). Then \(x \in \overline{A}\) if and only if for every \(\delta > 0\), \(B(x,\delta) \cap A \not=\emptyset\).

Let us prove the two contrapositives. Let us show that \(x \notin \overline{A}\) if and only if there exists a \(\delta > 0\) such that \(B(x,\delta) \cap A = \emptyset\).

First suppose \(x \notin \overline{A}\). We know \(\overline{A}\) is closed. Thus there is a \(\delta > 0\) such that \(B(x,\delta) \subset \overline{A}^c\). As \(A \subset \overline{A}\) we see that \(B(x,\delta) \subset A^c\) and hence \(B(x,\delta) \cap A = \emptyset\).

On the other hand suppose there is a \(\delta > 0\) such that \(B(x,\delta) \cap A = \emptyset\). Then \({B(x,\delta)}^c\) is a closed set and we have that \(A \subset {B(x,\delta)}^c\), but \(x \notin {B(x,\delta)}^c\). Thus as \(\overline{A}\) is the intersection of closed sets containing \(A\), we have \(x \notin \overline{A}\).

We can also talk about what is in the interior of a set and what is on the boundary.

Let \((X,d)\) be a metric space and \(A \subset X\), then the *interior* of \(A\) is the set \[A^\circ := \{ x \in A : \text{there exists a $\delta > 0$ such that $B(x,\delta) \subset A$} \} .\] The *boundary* of \(A\) is the set \[\partial A := \overline{A}\setminus A^\circ.\]

Suppose \(A=(0,1]\) and \(X = {\mathbb{R}}\). Then it is not hard to see that \(\overline{A}=[0,1]\), \(A^\circ = (0,1)\), and \(\partial A = \{ 0, 1 \}\).

Suppose \(X = \{ a, b \}\) with the discrete metric. Let \(A = \{ a \}\), then \(\overline{A} = A^\circ = A\) and \(\partial A = \emptyset\).

Let \((X,d)\) be a metric space and \(A \subset X\). Then \(A^\circ\) is open and \(\partial A\) is closed.

Given \(x \in A^\circ\) we have \(\delta > 0\) such that \(B(x,\delta) \subset A\). If \(z \in B(x,\delta)\), then as open balls are open, there is an \(\epsilon > 0\) such that \(B(z,\epsilon) \subset B(x,\delta) \subset A\), so \(z\) is in \(A^\circ\). Therefore \(B(x,\delta) \subset A^\circ\) and so \(A^\circ\) is open.

As \(A^\circ\) is open, then \(\partial A = \overline{A} \setminus A^\circ = \overline{A} \cap {(A^\circ)}^c\) is closed.

The boundary is the set of points that are close to both the set and its complement.

Let \((X,d)\) be a metric space and \(A \subset X\). Then \(x \in \partial A\) if and only if for every \(\delta > 0\), \(B(x,\delta) \cap A\) and \(B(x,\delta) \cap A^c\) are both nonempty.

Suppose \(x \in \partial A = \overline{A} \setminus A^\circ\) and let \(\delta > 0\) be arbitrary. By , \(B(x,\delta)\) contains a point from \(A\). If \(B(x,\delta)\) contained no points of \(A^c\), then \(x\) would be in \(A^\circ\). Hence \(B(x,\delta)\) contains a point of \(A^c\) as well.

Let us prove the other direction by contrapositive. If \(x \notin \overline{A}\), then there is some \(\delta > 0\) such that \(B(x,\delta) \subset \overline{A}^c\) as \(\overline{A}\) is closed. So \(B(x,\delta)\) contains no points of \(A\), because \(\overline{A}^c \subset A^c\).

Now suppose \(x \in A^\circ\), then there exists a \(\delta > 0\) such that \(B(x,\delta) \subset A\), but that means \(B(x,\delta)\) contains no points of \(A^c\).

We obtain the following immediate corollary about closures of \(A\) and \(A^c\). We simply apply .

Let \((X,d)\) be a metric space and \(A \subset X\). Then \(\partial A = \overline{A} \cap \overline{A^c}\).

## Exercises

Prove . Hint: consider the complements of the sets and apply .

Finish the proof of by proving that \(C(x,\delta)\) is closed.

Prove .

Suppose \((X,d)\) is a nonempty metric space with the discrete topology. Show that \(X\) is connected if and only if it contains exactly one element.

Show that if \(S \subset {\mathbb{R}}\) is a connected unbounded set, then it is an (unbounded) interval.

Show that every open set can be written as a union of closed sets.

a) Show that \(E\) is closed if and only if \(\partial E \subset E\). b) Show that \(U\) is open if and only if \(\partial U \cap U = \emptyset\).

a) Show that \(A\) is open if and only if \(A^\circ = A\). b) Suppose that \(U\) is an open set and \(U \subset A\). Show that \(U \subset A^\circ\).

Let \(X\) be a set and \(d\), \(d'\) be two metrics on \(X\). Suppose there exists an \(\alpha > 0\) and \(\beta > 0\) such that \(\alpha d(x,y) \leq d'(x,y) \leq \beta d(x,y)\) for all \(x,y \in X\). Show that \(U\) is open in \((X,d)\) if and only if \(U\) is open in \((X,d')\). That is, the topologies of \((X,d)\) and \((X,d')\) are the same.

Suppose \(\{ S_i \}\), \(i \in {\mathbb{N}}\) is a collection of connected subsets of a metric space \((X,d)\). Suppose there exists an \(x \in X\) such that \(x \in S_i\) for all \(i \in {\mathbb{N}}\). Show that \(\bigcup_{i=1}^\infty S_i\) is connected.

Let \(A\) be a connected set. a) Is \(\overline{A}\) connected? Prove or find a counterexample. b) Is \(A^\circ\) connected? Prove or find a counterexample. Hint: Think of sets in \({\mathbb{R}}^2\).

The definition of open sets in the following exercise is usually called the *subspace topology*. You are asked to show that we obtain the same topology by considering the subspace metric.

[exercise:mssubspace] Suppose \((X,d)\) is a metric space and \(Y \subset X\). Show that with the subspace metric on \(Y\), a set \(U \subset Y\) is open (in \(Y\)) whenever there exists an open set \(V \subset X\) such that \(U = V \cap Y\).

Let \((X,d)\) be a metric space. a) For any \(x \in X\) and \(\delta > 0\), show \(\overline{B(x,\delta)} \subset C(x,\delta)\). b) Is it always true that \(\overline{B(x,\delta)} = C(x,\delta)\)? Prove or find a counterexample.

Let \((X,d)\) be a metric space and \(A \subset X\). Show that \(A^\circ = \bigcup \{ V : V \subset A \text{ is open} \}\).

### Sequences and convergence

*Note: 1 lecture*

#### Sequences

The notion of a sequence in a metric space is very similar to a sequence of real numbers. The definitions are essentially the same as those for real numbers in the sense of where \({\mathbb{R}}\) with the standard metric \(d(x,y)=\left\lvert {x-y} \right\rvert\) is replaced by an arbitrary metric space \((X,d)\).

A *sequence* in a metric space \((X,d)\) is a function \(x \colon {\mathbb{N}}\to X\). As before we write \(x_n\) for the \(n\)th element in the sequence and use the notation \(\{ x_n \}\), or more precisely \[\{ x_n \}_{n=1}^\infty .\]

A sequence \(\{ x_n \}\) is *bounded* if there exists a point \(p \in X\) and \(B \in {\mathbb{R}}\) such that \[d(p,x_n) \leq B \qquad \text{for all $n \in {\mathbb{N}}$.}\] In other words, the sequence \(\{x_n\}\) is bounded whenever the set \(\{ x_n : n \in {\mathbb{N}}\}\) is bounded.

If \(\{ n_j \}_{j=1}^\infty\) is a sequence of natural numbers such that \(n_{j+1} > n_j\) for all \(j\), then the sequence \(\{ x_{n_j} \}_{j=1}^\infty\) is said to be a *subsequence* of \(\{x_n \}\).

Similarly we also define convergence. Again, we will be cheating a little bit and we will use the definite article in front of the word *limit* before we prove that the limit is unique.

A sequence \(\{ x_n \}\) in a metric space \((X,d)\) is said to *converge* to a point \(p \in X\), if for every \(\epsilon > 0\), there exists an \(M \in {\mathbb{N}}\) such that \(d(x_n,p) < \epsilon\) for all \(n \geq M\). The point \(p\) is said to be the *limit* of \(\{ x_n \}\). We write \[\lim_{n\to \infty} x_n := p .\]

A sequence that converges is said to be *convergent*. Otherwise, the sequence is said to be *divergent*.

Let us prove that the limit is unique. Note that the proof is almost identical to the proof of the same fact for sequences of real numbers. Many results we know for sequences of real numbers can be proved in the more general settings of metric spaces. We must replace \(\left\lvert {x-y} \right\rvert\) with \(d(x,y)\) in the proofs and apply the triangle inequality correctly.

[prop:mslimisunique] A convergent sequence in a metric space has a unique limit.

Suppose the sequence \(\{ x_n \}\) has the limit \(x\) and the limit \(y\). Take an arbitrary \(\epsilon > 0\). From the definition find an \(M_1\) such that for all \(n \geq M_1\), \(d(x_n,x) < \nicefrac{\epsilon}{2}\). Similarly find an \(M_2\) such that for all \(n \geq M_2\) we have \(d(x_n,y) < \nicefrac{\epsilon}{2}\). Now take an \(n\) such that \(n \geq M_1\) and also \(n \geq M_2\) \[\begin{split} d(y,x) & \leq d(y,x_n) + d(x_n,x) \\ & < \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon . \end{split}\] As \(d(y,x) < \epsilon\) for all \(\epsilon > 0\), then \(d(x,y) = 0\) and \(y=x\). Hence the limit (if it exists) is unique.

The proofs of the following propositions are left as exercises.

[prop:msconvbound] A convergent sequence in a metric space is bounded.

[prop:msconvifa] A sequence \(\{ x_n \}\) in a metric space \((X,d)\) converges to \(p \in X\) if and only if there exists a sequence \(\{ a_n \}\) of real numbers such that \[d(x_n,p) \leq a_n \quad \text{for all $n \in {\mathbb{N}}$},\] and \[\lim_{n\to\infty} a_n = 0.\]

[prop:mssubseq] Let \(\{ x_n \}\) be a sequence in a metric space \((X,d)\).

- If \(\{ x_n \}\) converges to \(p \in X\), then every subsequence \(\{ x_{n_k} \}\) converges to \(p\).
- If for some \(K \in {\mathbb{N}}\) the \(K\)-tail \(\{ x_n \}_{n=K+1}^\infty\) converges to \(p \in X\), then \(\{ x_n \}\) converges to \(p\).

#### Convergence in euclidean space

It is useful to note what convergence means in the euclidean space \({\mathbb{R}}^n\).

[prop:msconveuc] Let \(\{ x_j \}_{j=1}^\infty\) be a sequence in \({\mathbb{R}}^n\), where we write \(x_j = \bigl(x_{j,1},x_{j,2},\ldots,x_{j,n}\bigr) \in {\mathbb{R}}^n\). Then \(\{ x_j \}_{j=1}^\infty\) converges if and only if \(\{ x_{j,k} \}_{j=1}^\infty\) converges for every \(k\), in which case \[\lim_{j\to\infty} x_j = \Bigl( \lim_{j\to\infty} x_{j,1}, \lim_{j\to\infty} x_{j,2}, \ldots, \lim_{j\to\infty} x_{j,n} \Bigr) .\]

Let \(\{ x_j \}_{j=1}^\infty\) be a convergent sequence in \({\mathbb{R}}^n\), where we write \(x_j = \bigl(x_{j,1},x_{j,2},\ldots,x_{j,n}\bigr) \in {\mathbb{R}}^n\). Let \(y = (y_1,y_2,\ldots,y_n) \in {\mathbb{R}}^n\) be the limit. Given \(\epsilon > 0\), there exists an \(M\) such that for all \(j \geq M\) we have \[d(y,x_j) < \epsilon.\] Fix some \(k=1,2,\ldots,n\). For \(j \geq M\) we have \[\bigl\lvert y_k - x_{j,k} \bigr\rvert = \sqrt{{\bigl(y_k - x_{j,k} \bigr)}^2} \leq \sqrt{\sum_{\ell=1}^n {\bigl(y_\ell-x_{j,\ell}\bigr)}^2} = d(y,x_j) < \epsilon .\] Hence the sequence \(\{ x_{j,k} \}_{j=1}^\infty\) converges to \(y_k\).

For the other direction suppose \(\{ x_{j,k} \}_{j=1}^\infty\) converges to \(y_k\) for every \(k=1,2,\ldots,n\). Hence, given \(\epsilon > 0\), pick an \(M\), such that if \(j \geq M\) then \(\bigl\lvert y_k-x_{j,k} \bigr\rvert < \nicefrac{\epsilon}{\sqrt{n}}\) for all \(k=1,2,\ldots,n\). Then \[d(y,x_j) = \sqrt{\sum_{k=1}^n {\bigl(y_k-x_{j,k}\bigr)}^2} < \sqrt{\sum_{k=1}^n {\left(\frac{\epsilon}{\sqrt{n}}\right)}^2} = \sqrt{\sum_{k=1}^n \frac

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[40]/div/div[2]/p[4]/span, line 1, column 1 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

#### Convergence and topology

The topology, that is, the set of open sets of a space encodes which sequences converge.

[prop:msconvtopo] Let \((X,d)\) be a metric space and \(\{x_n\}\) a sequence in \(X\). Then \(\{ x_n \}\) converges to \(x \in X\) if and only if for every open neighborhood \(U\) of \(x\), there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \(x_n \in U\).

First suppose \(\{ x_n \}\) converges. Let \(U\) be an open neighborhood of \(x\), then there exists an \(\epsilon > 0\) such that \(B(x,\epsilon) \subset U\). As the sequence converges, find an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) we have \(d(x,x_n) < \epsilon\) or in other words \(x_n \in B(x,\epsilon) \subset U\).

Let us prove the other direction. Given \(\epsilon > 0\) let \(U := B(x,\epsilon)\) be the neighborhood of \(x\). Then there is an \(M \in {\mathbb{N}}\) such that for \(n \geq M\) we have \(x_n \in U = B(x,\epsilon)\) or in other words, \(d(x,x_n) < \epsilon\).

A set is closed when it contains the limits of its convergent sequences.

[prop:msclosedlim] Let \((X,d)\) be a metric space, \(E \subset X\) a closed set and \(\{ x_n \}\) a sequence in \(E\) that converges to some \(x \in X\). Then \(x \in E\).

Let us prove the contrapositive. Suppose \(\{ x_n \}\) is a sequence in \(X\) that converges to \(x \in E^c\). As \(E^c\) is open, says there is an \(M\) such that for all \(n \geq M\), \(x_n \in E^c\). So \(\{ x_n \}\) is not a sequence in \(E\).

When we take a closure of a set \(A\), we really throw in precisely those points that are limits of sequences in \(A\).

[prop:msclosureapprseq] Let \((X,d)\) be a metric space and \(A \subset X\). Then \(x \in \overline{A}\) if and only if there exists a sequence \(\{ x_n \}\) of elements in \(A\) such that \(\lim\, x_n = x\).

Let \(x \in \overline{A}\). We know by that given \(\nicefrac{1}{n}\), there exists a point \(x_n \in B(x,\nicefrac{1}{n}) \cap A\). As \(d(x,x_n) < \nicefrac{1}{n}\), we have \(\lim\, x_n = x\).

For the other direction, see .

## Exercises

[exercise:reverseclosedseq] Let \((X,d)\) be a metric space and let \(A \subset X\). Let \(E\) be the set of all \(x \in X\) such that there exists a sequence \(\{ x_n \}\) in \(A\) that converges to \(x\). Show \(E = \overline{A}\).

a) Show that \(d(x,y) := \min \{ 1, \left\lvert {x-y} \right\rvert \}\) defines a metric on \({\mathbb{R}}\). b) Show that a sequence converges in \(({\mathbb{R}},d)\) if and only if it converges in the standard metric. c) Find a bounded sequence in \(({\mathbb{R}},d)\) that contains no convergent subsequence.

Prove .

Prove .

Suppose \(\{x_n\}_{n=1}^\infty\) converges to \(x\). Suppose \(f \colon {\mathbb{N}} \to {\mathbb{N}}\) is a one-to-one function. Show that \(\{ x_{f(n)} \}_{n=1}^\infty\) converges to \(x\).

If \((X,d)\) is a metric space where \(d\) is the discrete metric. Suppose \(\{ x_n \}\) is a convergent sequence in \(X\). Show that there exists a \(K \in {\mathbb{N}}\) such that for all \(n \geq K\) we have \(x_n = x_K\).

A set \(S \subset X\) is said to be dense in \(X\) if for every \(x \in X\), there exists a sequence \(\{ x_n \}\) in \(S\) that converges to \(x\). Prove that \({\mathbb{R}}^n\) contains a countable dense subset.

Suppose \(\{ U_n \}_{n=1}^\infty\) be a decreasing (\(U_{n+1} \subset U_n\) for all \(n\)) sequence of open sets in a metric space \((X,d)\) such that \(\bigcap_{n=1}^\infty U_n = \{ p \}\) for some \(p \in X\). Suppose \(\{ x_n \}\) is a sequence of points in \(X\) such that \(x_n \in U_n\). Does \(\{ x_n \}\) necessarily converge to \(p\)? Prove or construct a counterexample.

Let \(E \subset X\) be closed and let \(\{ x_n \}\) be a sequence in \(X\) converging to \(p \in X\). Suppose \(x_n \in E\) for infinitely many \(n \in {\mathbb{N}}\). Show \(p \in E\).

Take \({\mathbb{R}}^* = \{ -\infty \} \cup {\mathbb{R}}\cup \{ \infty \}\) be the extended reals. Define \(d(x,y) := \bigl\lvert \frac{x}{1+\left\lvert {x} \right\rvert} - \frac{y}{1+\left\lvert {y} \right\rvert} \bigr\rvert\) if \(x, y \in {\mathbb{R}}\), define \(d(\infty,x) := \bigl\lvert 1 - \frac{x}{1+\left\lvert {x} \right\rvert} \bigr\rvert\), \(d(-\infty,x) := \bigl\lvert 1 + \frac{x}{1+\left\lvert {x} \right\rvert} \bigr\rvert\) for all \(x \in {\mathbb{R}}\), and let \(d(\infty,-\infty) := 2\). a) Show that \(({\mathbb{R}}^*,d)\) is a metric space. b) Suppose \(\{ x_n \}\) is a sequence of real numbers such that for every \(M \in {\mathbb{R}}\), there exists an \(N\) such that \(x_n \geq M\) for all \(n \geq N\). Show that \(\lim\, x_n = \infty\) in \(({\mathbb{R}}^*,d)\). c) Show that a sequence of real numbers converges to a real number in \(({\mathbb{R}}^*,d)\) if and only if it converges in \({\mathbb{R}}\) with the standard metric.

Suppose \(\{ V_n \}_{n=1}^\infty\) is a collection of open sets in \((X,d)\) such that \(V_{n+1} \supset V_n\). Let \(\{ x_n \}\) be a sequence such that \(x_n \in V_{n+1} \setminus V_n\) and suppose \(\{ x_n \}\) converges to \(p \in X\). Show that \(p \in \partial V\) where \(V = \bigcup_{n=1}^\infty V_n\).

Prove .

### Completeness and compactness

*Note: 2 lectures*

#### Cauchy sequences and completeness

Just like with sequences of real numbers we define Cauchy sequences.

Let \((X,d)\) be a metric space. A sequence \(\{ x_n \}\) in \(X\) is a *Cauchy sequence* if for every \(\epsilon > 0\) there exists an \(M \in {\mathbb{N}}\) such that for all \(n \geq M\) and all \(k \geq M\) we have \[d(x_n, x_k) < \epsilon .\]

The definition is again simply a translation of the concept from the real numbers to metric spaces. So a sequence of real numbers is Cauchy in the sense of if and only if it is Cauchy in the sense above, provided we equip the real numbers with the standard metric \(d(x,y) = \left\lvert {x-y} \right\rvert\).

A convergent sequence in a metric space is Cauchy.

Suppose \(\{ x_n \}\) converges to \(x\). Given \(\epsilon > 0\) there is an \(M\) such that for \(n \geq M\) we have \(d(x,x_n) < \nicefrac{\epsilon}{2}\). Hence for all \(n,k \geq M\) we have \(d(x_n,x_k) \leq d(x_n,x) + d(x,x_k) < \nicefrac{\epsilon}{2} + \nicefrac{\epsilon}{2} = \epsilon\).

Let \((X,d)\) be a metric space. We say \(X\) is *complete* or *Cauchy-complete* if every Cauchy sequence \(\{ x_n \}\) in \(X\) converges to an \(x \in X\).

The space \({\mathbb{R}}^n\) with the standard metric is a complete metric space.

For \({\mathbb{R}}= {\mathbb{R}}^1\) completeness was proved in . The proof of the above proposition is a reduction to the one dimensional case.

Let \(\{ x_j \}_{j=1}^\infty\) be a Cauchy sequence in \({\mathbb{R}}^n\), where we write \(x_j = \bigl(x_{j,1},x_{j,2},\ldots,x_{j,n}\bigr) \in {\mathbb{R}}^n\). As the sequence is Cauchy, given \(\epsilon > 0\), there exists an \(M\) such that for all \(i,j \geq M\) we have \[d(x_i,x_j) < \epsilon.\]

Fix some \(k=1,2,\ldots,n\), for \(i,j \geq M\) we have \[\bigl\lvert x_{i,k} - x_{j,k} \bigr\rvert = \sqrt

Callstack: at (Bookshelves/Analysis/Introduction_to_Real_Analysis_(Lebl)/02:_Real_Numbers/2.01:_Basic_properties), /content/body/div[41]/div/div[1]/p[10]/span, line 1, column 1 at wiki.page() at (Under_Construction/Purgatory/Remixer_University/Username:_junalyn2020/Book:_Introduction_to_Real_Analysis_(Lebl)/1:_Real_Numbers/1.1:_Basic_properties), /content/body/div[1]/pre, line 2, column 14

Write \(y = (y_1,y_2,\ldots,y_n) \in {\mathbb{R}}^n\). By we have that \(\{ x_j \}\) converges to \(y \in {\mathbb{R}}^n\) and hence \({\mathbb{R}}^n\) is complete.

Note that a subset of \({\mathbb{R}}^n\) with the subspace metric need not be complete. For example, \((0,1]\) with the subspace metric is not complete as \(\{ \nicefrac{1}{n} \}\) is a Cauchy sequence in \((0,1]\) with no limit in \((0,1]\). But see also .

#### Compactness

Let \((X,d)\) be a metric space and \(K \subset X\). The set \(K\) is said to be *compact* if for any collection of open sets \(\{ U_{\lambda} \}_{\lambda \in I}\) such that \[K \subset \bigcup_{\lambda \in I} U_\lambda ,\] there exists a finite subset \(\{ \lambda_1, \lambda_2,\ldots,\lambda_k \} \subset I\) such that \[K \subset \bigcup_{j=1}^k U_{\lambda_j} .\]

A collection of open sets \(\{ U_{\lambda} \}_{\lambda \in I}\) as above is said to be a *open cover* of \(K\). So a way to say that \(K\) is compact is to say that *every open cover of \(K\) has a finite subcover*.

Let \((X,d)\) be a metric space. A compact set \(K \subset X\) is closed and bounded.

First, we prove that a compact set is bounded. Fix \(p \in X\). We have the open cover \[K \subset \bigcup_{n=1}^\infty B(p,n) = X .\] If \(K\) is compact, then there exists some set of indices \(n_1 < n_2 < \ldots < n_k\) such that \[K \subset \bigcup_{j=1}^k B(p,n_j) = B(p,n_k) .\] As \(K\) is contained in a ball, \(K\) is bounded.

Next, we show a set that is not closed is not compact. Suppose \(\overline{K} \not= K\), that is, there is a point \(x \in \overline{K} \setminus K\). If \(y \not= x\), then for \(n\) with \(\nicefrac{1}{n} < d(x,y)\) we have \(y \notin C(x,\nicefrac{1}{n})\). Furthermore \(x \notin K\), so \[K \subset \bigcup_{n=1}^\infty {C(x,\nicefrac{1}{n})}^c .\] As a closed ball is closed, \({C(x,\nicefrac{1}{n})}^c\) is open, and so we have an open cover. If we take any finite collection of indices \(n_1 < n_2 < \ldots < n_k\), then \[\bigcup_{j=1}^k {C(x,\nicefrac{1}{n_j})}^c = {C(x,\nicefrac{1}{n_k})}^c\] As \(x\) is in the closure, \(C(x,\nicefrac{1}{n_k}) \cap K \not= \emptyset\). So there is no finite subcover and \(K\) is not compact.

We prove below that in finite dimensional euclidean space every closed bounded set is compact. So closed bounded sets of \({\mathbb{R}}^n\) are examples of compact sets. It is not true that in every metric space, closed and bounded is equivalent to compact. A simple example would be an incomplete metric space such as \((0,1)\) with the subspace metric. But there are many complete and very useful metric spaces where closed and bounded is not enough to give compactness, see : \(C([a,b],{\mathbb{R}})\) is a complete metric space, but the closed unit ball \(C(0,1)\) is not compact. However, see .

A useful property of compact sets in a metric space is that every sequence has a convergent subsequence. Such sets are sometimes called *sequentially compact*. Let us prove that in the context of metric spaces, a set is compact if and only if it is sequentially compact. First we prove a lemma.

[ms:lebesgue] Let \((X,d)\) be a metric space and \(K \subset X\). Suppose every sequence in \(K\) has a subsequence convergent in \(K\). Given an open cover \(\{ U_\lambda \}_{\lambda \in I}\) of \(K\), there exists a \(\delta > 0\) such that for every \(x \in K\), there exists a \(\lambda \in I\) with \(B(x,\delta) \subset U_\lambda\).

It is important to recognize what the lemma says. It says that given any cover there is a single \(\delta > 0\). The \(\delta\) can depend on the cover, but of course it does not depend on \(x\).

Let us prove the lemma by contrapositive. If the conclusion is not true, then there is an open cover \(\{ U_\lambda \}_{\lambda \in I}\) of \(K\) with the following property. For every \(n \in {\mathbb{N}}\) there exists an \(x_n \in K\) such that \(B(x_n,\nicefrac{1}{n})\) is not a subset of any \(U_\lambda\). Given any \(x \in K\), there is a \(\lambda \in I\) such that \(x \in U_\lambda\). Hence there is an \(\epsilon > 0\) such that \(B(x,\epsilon) \subset U_\lambda\). Take \(M\) such that \(\nicefrac{1}{M} < \nicefrac{\epsilon}{2}\). If \(y \in B(x,\nicefrac{\epsilon}{2})\) and \(n \geq M\), then by triangle inequality \[B(y,\nicefrac{1}{n}) \subset B(y,\nicefrac{1}{M}) \subset B(y,\nicefrac{\epsilon}{2}) \subset B(x,\epsilon) \subset U_\lambda .\] In other words, for all \(n \geq M\), \(x_n \notin B(x,\nicefrac{\epsilon}{2})\). Hence the sequence cannot have a subsequence converging to \(x\). As \(x \in K\) was arbitrary we are done.

[thm:mscompactisseqcpt] Let \((X,d)\) be a metric space. Then \(K \subset X\) is a compact set if and only if every sequence in \(K\) has a subsequence converging to a point in \(K\).

Let \(K \subset X\) be a set and \(\{ x_n \}\) a sequence in \(K\). Suppose that for each \(x \in K\), there is a ball \(B(x,\alpha_x)\) for some \(\alpha_x > 0\) such that \(x_n \in B(x,\alpha_x)\) for only finitely many \(n \in {\mathbb{N}}\). Then \[K \subset \bigcup_{x \in K} B(x,\alpha_x) .\] Any finite collection of these balls is going to contain only finitely many \(x_n\). Thus for any finite collection of such balls there is an \(x_n \in K\) that is not in the union. Therefore, \(K\) is not compact.

So if \(K\) is compact, then there exists an \(x \in K\) such that for any \(\delta > 0\), \(B(x,\delta)\) contains \(x_k\) for infinitely many \(k \in {\mathbb{N}}\). The ball \(B(x,1)\) contains some \(x_k\) so let \(n_1 := k\). If \(n_{j-1}\) is defined, then there must exist a \(k > n_{j-1}\) such that \(x_k \in B(x,\nicefrac{1}{j})\), so define \(n_j := k\). Notice that \(d(x,x_{n_j}) < \nicefrac{1}{j}\). By , \(\lim\, x_{n_j} = x\).

For the other direction, suppose every sequence in \(K\) has a subsequence converging in \(K\). Take an open cover \(\{ U_\lambda \}_{\lambda \in I}\) of \(K\). Using the Lebesgue covering lemma above, we find a \(\delta > 0\) such that for every \(x\), there is a \(\lambda \in I\) with \(B(x,\delta) \subset U_\lambda\).

Pick \(x_1 \in K\) and find \(\lambda_1 \in I\) such that \(B(x_1,\delta) \subset U_{\lambda_1}\). If \(K \subset U_{\lambda_1}\), we stop as we have found a finite subcover. Otherwise, there must be a point \(x_2 \in K \setminus U_{\lambda_1}\). Note that \(d(x_2,x_1) \geq \delta\). There must exist some \(\lambda_2 \in I\) such that \(B(x_2,\delta) \subset U_{\lambda_2}\). We work inductively. Suppose \(\lambda_{n-1}\) is defined. Either \(U_{\lambda_1} \cup U_{\lambda_2} \cup \cdots \cup U_{\lambda_{n-1}}\) is a finite cover of \(K\), in which case we stop, or there must be a point \(x_n \in K \setminus \bigl( U_{\lambda_1} \cup U_{\lambda_2} \cup \cdots \cup U_{\lambda_{n-1}}\bigr)\). Note that \(d(x_n,x_j) \geq \delta\) for all \(j = 1,2,\ldots,n-1\). Next, there must be some \(\lambda_n \in I\) such that \(B(x_n,\delta) \subset U_{\lambda_n}\).

Either at some point we obtain a finite subcover of \(K\) or we obtain an infinite sequence \(\{ x_n \}\) as above. For contradiction suppose that there is no finite subcover and we have the sequence \(\{ x_n \}\). For all \(n\) and \(k\), \(n \not= k\), we have \(d(x_n,x_k) \geq \delta\), so no subsequence of \(\{ x_n \}\) can be Cauchy. Hence no subsequence of \(\{ x_n \}\) can be convergent, which is a contradiction.

The Bolzano-Weierstrass theorem for sequences of real numbers () says that any bounded sequence in \({\mathbb{R}}\) has a convergent subsequence. Therefore any sequence in a closed interval \([a,b] \subset {\mathbb{R}}\) has a convergent subsequence. The limit must also be in \([a,b]\) as limits preserve non-strict inequalities. Hence a closed bounded interval \([a,b] \subset {\mathbb{R}}\) is compact.

Let \((X,d)\) be a metric space and let \(K \subset X\) be compact. If \(E \subset K\) is a closed set, then \(E\) is compact.

Let \(\{ x_n \}\) be a sequence in \(E\). It is also a sequence in \(K\). Therefore it has a convergent subsequence \(\{ x_{n_j} \}\) that converges to some \(x \in K\). As \(E\) is closed the limit of a sequence in \(E\) is also in \(E\) and so \(x \in E\). Thus \(E\) must be compact.

[thm:msbw] A closed bounded subset \(K \subset {\mathbb{R}}^n\) is compact.

So subsets of \({\mathbb{R}}^n\) are compact if and only if they are closed and bounded, a condition that is much easier to check. Let us reiterate that the Heine-Borel theorem only holds for \({\mathbb{R}}^n\) and not for metric spaces in general. In general, compact implies closed and bounded, but not vice versa.

For \({\mathbb{R}}= {\mathbb{R}}^1\) if \(K \subset {\mathbb{R}}\) is closed and bounded, then any sequence \(\{ x_k \}\) in \(K\) is bounded, so it has a convergent subsequence by Bolzano-Weierstrass theorem (). As \(K\) is closed, the limit of the subsequence must be an element of \(K\). So \(K\) is compact.

Let us carry out the proof for \(n=2\) and leave arbitrary \(n\) as an exercise. As \(K \subset {\mathbb{R}}^2\) is bounded, there exists a set \(B=[a,b]\times[c,d] \subset {\mathbb{R}}^2\) such that \(K \subset B\). We will show that \(B\) is compact. Then \(K\), being a closed subset of a compact \(B\), is also compact.

Let \(\{ (x_k,y_k) \}_{k=1}^\infty\) be a sequence in \(B\). That is, \(a \leq x_k \leq b\) and \(c \leq y_k \leq d\) for all \(k\). A bounded sequence of real numbers has a convergent subsequence so there is a subsequence \(\{ x_{k_j} \}_{j=1}^\infty\) that is convergent. The subsequence \(\{ y_{k_j} \}_{j=1}^\infty\) is also a bounded sequence so there exists a subsequence \(\{ y_{k_{j_i}} \}_{i=1}^\infty\) that is convergent. A subsequence of a convergent sequence is still convergent, so \(\{ x_{k_{j_i}} \}_{i=1}^\infty\) is convergent. Let \[x := \lim_{i\to\infty} x_{k_{j_i}} \qquad \text{and} \qquad y := \lim_{i\to\infty} y_{k_{j_i}} .\] By , \(\bigl\{ (x_{k_{j_i}},y_{k_{j_i}}) \bigr\}_{i=1}^\infty\) converges to \((x,y)\). Furthermore, as \(a \leq x_k \leq b\) and \(c \leq y_k \leq d\) for all \(k\), we know that \((x,y) \in B\).

The discrete metric provides interesting counterexamples again. Let \((X,d)\) be a metric space with the discrete metric, that is \(d(x,y) = 1\) if \(x \not= y\). Suppose \(X\) is an infinite set. Then:

- \((X,d)\) is a complete metric space.
- Any subset \(K \subset X\) is closed and bounded.
- A subset \(K \subset X\) is compact if and only if it is a finite set.
- The conclusion of the Lebesgue covering lemma is always satisfied with e.g. \(\delta = \nicefrac{1}{2}\), even for noncompact \(K \subset X\).

The proofs of these statements are either trivial or are relegated to the exercises below.

## Exercises

Let \((X,d)\) be a metric space and \(A\) a finite subset of \(X\). Show that \(A\) is compact.

Let \(A = \{ \nicefrac{1}{n} : n \in {\mathbb{N}}\} \subset {\mathbb{R}}\). a) Show that \(A\) is not compact directly using the definition. b) Show that \(A \cup \{ 0 \}\) is compact directly using the definition.

Let \((X,d)\) be a metric space with the discrete metric. a) Prove that \(X\) is complete. b) Prove that \(X\) is compact if and only if \(X\) is a finite set.

a) Show that the union of finitely many compact sets is a compact set. b) Find an example where the union of infinitely many compact sets is not compact.

Prove for arbitrary dimension. Hint: The trick is to use the correct notation.

Show that a compact set \(K\) is a complete metric space (using the subspace metric).

[exercise:CabRcomplete] Let \(C([a,b],{\mathbb{R}})\) be the metric space as in . Show that \(C([a,b],{\mathbb{R}})\) is a complete metric space.

[exercise:msclbounnotcompt] Let \(C([0,1],{\mathbb{R}})\) be the metric space of . Let \(0\) denote the zero function. Then show that the closed ball \(C(0,1)\) is not compact (even though it is closed and bounded). Hints: Construct a sequence of distinct continuous functions \(\{ f_n \}\) such that \(d(f_n,0) = 1\) and \(d(f_n,f_k) = 1\) for all \(n \not= k\). Show that the set \(\{ f_n : n \in {\mathbb{N}}\} \subset C(0,1)\) is closed but not compact. See for inspiration.

Show that there exists a metric on \({\mathbb{R}}\) that makes \({\mathbb{R}}\) into a compact set.

Suppose \((X,d)\) is complete and suppose we have a countably infinite collection of nonempty compact sets \(E_1 \supset E_2 \supset E_3 \supset \cdots\) then prove \(\bigcap_{j=1}^\infty E_j \not= \emptyset\).

Let \(C([0,1],{\mathbb{R}})\) be the metric space of . Let \(K\) be the set of \(f \in C([0,1],{\mathbb{R}})\) such that \(f\) is equal to a quadratic polynomial, i.e. \(f(x) = a+bx+cx^2\), and such that \(\left\lvert {f(x)} \right\rvert \leq 1\) for all \(x \in [0,1]\), that is \(f \in C(0,1)\). Show that \(K\) is compact.

[exercise:mstotbound] Let \((X,d)\) be a complete metric space. Show that \(K \subset X\) is compact if and only if \(K\) is closed and such that for every \(\epsilon > 0\) there exists a finite set of points \(x_1,x_2,\ldots,x_n\) with \(K \subset \bigcup_{j=1}^n B(x_j,\epsilon)\). Note: Such a set \(K\) is said to be *totally bounded*, so in a complete metric space a set is compact if and only if it is closed and totally bounded.

Take \({\mathbb{N}}\subset {\mathbb{R}}\) using the standard metric. Find an open cover of \({\mathbb{N}}\) such that the conclusion of the Lebesgue covering lemma does not hold.

Prove the general Bolzano-Weierstrass theorem: Any bounded sequence \(\{ x_k \}\) in \({\mathbb{R}}^n\) has a convergent subsequence.

Let \(X\) be a metric space and \(C \subset {\mathcal{P}}(X)\) the set of nonempty compact subsets of \(X\). Using the Hausdorff metric from , show that \((C,d_H)\) is a metric space. That is, show that if \(L\) and \(K\) are nonempty compact subsets then \(d_H(L,K) = 0\) if and only if \(L=K\).

[exercise:closedcomplete] Let \((X,d)\) be a complete metric space and \(E \subset X\) a closed set. Show that \(E\) with the subspace metric is a complete metric space.

### Continuous functions

*Note: 1 lecture*

#### Continuity

Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces and \(c \in X\). Then \(f \colon X \to Y\) is *continuous at \(c\)* if for every \(\epsilon > 0\) there is a \(\delta > 0\) such that whenever \(x \in X\) and \(d_X(x,c) < \delta\), then \(d_Y\bigl(f(x),f(c)\bigr) < \epsilon\).

When \(f \colon X \to Y\) is continuous at all \(c \in X\), then we simply say that \(f\) is a *continuous function*.

The definition agrees with the definition from when \(f\) is a real-valued function on the real line, if we take the standard metric on \({\mathbb{R}}\).

[prop:contiscont] Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces. Then \(f \colon X \to Y\) is continuous at \(c \in X\) if and only if for every sequence \(\{ x_n \}\) in \(X\) converging to \(c\), the sequence \(\{ f(x_n) \}\) converges to \(f(c)\).

Suppose \(f\) is continuous at \(c\). Let \(\{ x_n \}\) be a sequence in \(X\) converging to \(c\). Given \(\epsilon > 0\), there is a \(\delta > 0\) such that \(d_X(x,c) < \delta\) implies \(d_Y\bigl(f(x),f(c)\bigr) < \epsilon\). So take \(M\) such that for all \(n \geq M\), we have \(d_X(x_n,c) < \delta\), then \(d_Y\bigl(f(x_n),f(c)\bigr) < \epsilon\). Hence \(\{ f(x_n) \}\) converges to \(f(c)\).

On the other hand suppose \(f\) is not continuous at \(c\). Then there exists an \(\epsilon > 0\), such that for every \(n \in {\mathbb{N}}\) there exists an \(x_n \in X\), with \(d_X(x_n,c) < \nicefrac{1}{n}\) such that \(d_Y\bigl(f(x_n),f(c)\bigr) \geq \epsilon\). Then \(\{ x_n \}\) converges to \(c\), but \(\{ f(x_n) \}\) does not converge to \(f(c)\).

Suppose \(f \colon {\mathbb{R}}^2 \to {\mathbb{R}}\) is a polynomial. That is, \[f(x,y) = \sum_{j=0}^d \sum_{k=0}^{d-j} a_{jk}\,x^jy^k = a_{0\,0} + a_{1\,0} \, x + a_{0\,1} \, y+ a_{2\,0} \, x^2+ a_{1\,1} \, xy+ a_{0\,2} \, y^2+ \cdots + a_{0\,d} \, y^d ,\] for some \(d \in {\mathbb{N}}\) (the degree) and \(a_{jk} \in {\mathbb{R}}\). Then we claim \(f\) is continuous. Let \(\{ (x_n,y_n) \}_{n=1}^\infty\) be a sequence in \({\mathbb{R}}^2\) that converges to \((x,y) \in {\mathbb{R}}^2\). We have proved that this means that \(\lim\, x_n = x\) and \(\lim\, y_n = y\). So by we have \[\lim_{n\to\infty} f(x_n,y_n) = \lim_{n\to\infty} \sum_{j=0}^d \sum_{k=0}^{d-j} a_{jk} \, x_n^jy_n^k = \sum_{j=0}^d \sum_{k=0}^{d-j} a_{jk} \, x^jy^k = f(x,y) .\] So \(f\) is continuous at \((x,y)\), and as \((x,y)\) was arbitrary \(f\) is continuous everywhere. Similarly, a polynomial in \(n\) variables is continuous.

#### Compactness and continuity

Continuous maps do not map closed sets to closed sets. For example, \(f \colon (0,1) \to {\mathbb{R}}\) defined by \(f(x) := x\) takes the set \((0,1)\), which is closed in \((0,1)\), to the set \((0,1)\), which is not closed in \({\mathbb{R}}\). On the other hand continuous maps do preserve compact sets.

[lemma:continuouscompact] Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces and \(f \colon X \to Y\) a continuous function. If \(K \subset X\) is a compact set, then \(f(K)\) is a compact set.

A sequence in \(f(K)\) can be written as \(\{ f(x_n) \}_{n=1}^\infty\), where \(\{ x_n \}_{n=1}^\infty\) is a sequence in \(K\). The set \(K\) is compact and therefore there is a subsequence \(\{ x_{n_i} \}_{i=1}^\infty\) that converges to some \(x \in K\). By continuity, \[\lim_{i\to\infty} f(x_{n_i}) = f(x) \in f(K) .\] So every sequence in \(f(K)\) has a subsequence convergent to a point in \(f(K)\), and \(f(K)\) is compact by .

As before, \(f \colon X \to {\mathbb{R}}\) achieves an *absolute minimum* at \(c \in X\) if \[f(x) \geq f(c) \qquad \text{ for all $x \in X$.}\] On the other hand, \(f\) achieves an *absolute maximum* at \(c \in X\) if \[f(x) \leq f(c) \qquad \text{ for all $x \in X$.}\]

Let \((X,d)\) be a compact metric space and \(f \colon X \to {\mathbb{R}}\) a continuous function. Then \(f\) is bounded and in fact \(f\) achieves an absolute minimum and an absolute maximum on \(X\).

As \(X\) is compact and \(f\) is continuous, we have that \(f(X) \subset {\mathbb{R}}\) is compact. Hence \(f(X)\) is closed and bounded. In particular, \(\sup f(X) \in f(X)\) and \(\inf f(X) \in f(X)\), because both the sup and inf can be achieved by sequences in \(f(X)\) and \(f(X)\) is closed. Therefore there is some \(x \in X\) such that \(f(x) = \sup f(X)\) and some \(y \in X\) such that \(f(y) = \inf f(X)\).

#### Continuity and topology

Let us see how to define continuity in terms of the topology, that is, the open sets. We have already seen that topology determines which sequences converge, and so it is no wonder that the topology also determines continuity of functions.

[lemma:mstopocontloc] Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces. A function \(f \colon X \to Y\) is continuous at \(c \in X\) if and only if for every open neighborhood \(U\) of \(f(c)\) in \(Y\), the set \(f^{-1}(U)\) contains an open neighborhood of \(c\) in \(X\).

First suppose that \(f\) is continuous at \(c\). Let \(U\) be an open neighborhood of \(f(c)\) in \(Y\), then \(B_Y\bigl(f(c),\epsilon\bigr) \subset U\) for some \(\epsilon > 0\). By continuity of \(f\), there exists a \(\delta > 0\) such that whenever \(x\) is such that \(d_X(x,c) < \delta\), then \(d_Y\bigl(f(x),f(c)\bigr) < \epsilon\). In other words, \[B_X(c,\delta) \subset f^{-1}\bigl(B_Y\bigl(f(c),\epsilon\bigr)\bigr) \subset f^{-1}(U) ,\] and \(B_X(c,\delta)\) is an open neighborhood of \(c\).

For the other direction, let \(\epsilon > 0\) be given. If \(f^{-1}\bigl(B_Y\bigl(f(c),\epsilon\bigr)\bigr)\) contains an open neighborhood \(W\) of \(c\), it contains a ball. That is, there is some \(\delta > 0\) such that \[B_X(c,\delta) \subset W \subset f^{-1}\bigl(B_Y\bigl(f(c),\epsilon\bigr)\bigr) .\] That means precisely that if \(d_X(x,c) < \delta\) then \(d_Y\bigl(f(x),f(c)\bigr) < \epsilon\), and so \(f\) is continuous at \(c\).

[thm:mstopocont] Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces. A function \(f \colon X \to Y\) is continuous if and only if for every open \(U \subset Y\), \(f^{-1}(U)\) is open in \(X\).

The proof follows from and is left as an exercise.

Let \(f \colon X \to Y\) be a continuous function. tells us that if \(E \subset Y\) is closed, then \(f^{-1}(E) = X \setminus f^{-1}(E^c)\) is also closed. Therefore if we have a continuous function \(f \colon X \to {\mathbb{R}}\), then the *zero set* of \(f\), that is, \(f^{-1}(0) = \{ x \in X : f(x) = 0 \}\), is closed. We have just proved the most basic result in *algebraic geometry*, the study of zero sets of polynomials.

Similarly the set where \(f\) is nonnegative, that is, \(f^{-1}\bigl( [0,\infty) \bigr) = \{ x \in X : f(x) \geq 0 \}\) is closed. On the other hand the set where \(f\) is positive, \(f^{-1}\bigl( (0,\infty) \bigr) = \{ x \in X : f(x) > 0 \}\) is open.

#### Uniform continuity

As for continuous functions on the real line, in the definition of continuity it is sometimes convenient to be able to pick one \(\delta\) for all points.

Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces. Then \(f \colon X \to Y\) is *uniformly continuous* if for every \(\epsilon > 0\) there is a \(\delta > 0\) such that whenever \(x,c \in X\) and \(d_X(x,c) < \delta\), then \(d_Y\bigl(f(x),f(c)\bigr) < \epsilon\).

A uniformly continuous function is continuous, but not necessarily vice-versa as we have seen.

[thm:Xcompactfunifcont] Let \((X,d_X)\) and \((Y,d_Y)\) be metric spaces. Suppose \(f \colon X \to Y\) is continuous and \(X\) compact. Then \(f\) is uniformly continuous.

Let \(\epsilon > 0\) be given. For each \(c \in X\), pick \(\delta_c > 0\) such that \(d_Y\bigl(f(x),f(c)\bigr) < \nicefrac{\epsilon}{2}\) whenever \(d_X(x,c) < \delta_c\). The balls \(B(c,\delta_c)\) cover \(X\), and the space \(X\) is compact. Apply the to obtain a \(\delta > 0\) such that for every \(x \in X\), there is a \(c \in X\) for which \(B(x,\delta) \subset B(c,\delta_c)\).

If \(x_1, x_2 \in X\) where \(d_X(x_1,x_2) < \delta\), find a \(c \in X\) such that \(B(x_1,\delta) \subset B(c,\delta_c)\). Then \(x_2 \in B(c,\delta_c)\). By the triangle inequality and the definition of \(\delta_c\) we have \[d_Y\bigl(f(x_1),f(x_2)\bigr) \leq d_Y\bigl(f(x_1),f(c)\bigr) + d_Y\bigl(f(c),f(x_2)\bigr) < \nicefrac{\epsilon}{2}+ \nicefrac{\epsilon}{2} = \epsilon . \qedhere\]

Useful examples of uniformly continuous functions are again the so-called *Lipschitz continuous* functions. That is if \((X,d_X)\) and \((Y,d_Y)\) are metric spaces, then \(f \colon X \to Y\) is called Lipschitz or \(K\)-Lipschitz if there exists a \(K \in {\mathbb{R}}\) such that \[d_Y\bigl(f(x),f(c)\bigr) \leq K d_X(x,c) \ \ \ \ \text{for all } x,c \in X.\] It is not difficult to prove that Lipschitz implies uniformly continuous, just take \(\delta = \nicefrac{\epsilon}{K}\). And we already saw in the case of functions on the real line, a function can be uniformly continuous but not Lipschitz.

It is worth mentioning that, if a function is Lipschitz, it tends to be easiest to simply show it is Lipschitz even if we are only interested in knowing continuity.

## Exercises

Consider \({\mathbb{N}}\subset {\mathbb{R}}\) with the standard metric. Let \((X,d)\) be a metric space and \(f \colon X \to {\mathbb{N}}\) a continuous function. a) Prove that if \(X\) is connected, then \(f\) is constant (the range of \(f\) is a single value). b) Find an example where \(X\) is disconnected and \(f\) is not constant.

Let \(f \colon {\mathbb{R}}^2 \to {\mathbb{R}}\) be defined by \(f(0,0) := 0\), and \(f(x,y) := \frac{xy}{x^2+y^2}\) if \((x,y) \not= (0,0)\). a) Show that for any fixed \(x\), the function that takes \(y\) to \(f(x,y)\) is continuous. Similarly for any fixed \(y\), the function that takes \(x\) to \(f(x,y)\) is continuous. b) Show that \(f\) is not continuous.

Suppose that \(f \colon X \to Y\) is continuous for metric spaces \((X,d_X)\) and \((Y,d_Y)\). Let \(A \subset X\). a) Show that \(f(\overline{A}) \subset \overline{f(A)}\). b) Show that the subset can be proper.

Prove . Hint: Use .

[exercise:msconnconn] Suppose \(f \colon X \to Y\) is continuous for metric spaces \((X,d_X)\) and \((Y,d_Y)\). Show that if \(X\) is connected, then \(f(X)\) is connected.

Prove the following version of the . Let \((X,d)\) be a connected metric space and \(f \colon X \to {\mathbb{R}}\) a continuous function. Suppose that there exist \(x_0,x_1 \in X\) and \(y \in {\mathbb{R}}\) such that \(f(x_0) < y < f(x_1)\). Then prove that there exists a \(z \in X\) such that \(f(z) = y\). Hint: See .

A continuous function \(f \colon X \to Y\) for metric spaces \((X,d_X)\) and \((Y,d_Y)\) is said to be *proper* if for every compact set \(K \subset Y\), the set \(f^{-1}(K)\) is compact. Suppose a continuous \(f \colon (0,1) \to (0,1)\) is proper and \(\{ x_n \}\) is a sequence in \((0,1)\) that converges to \(0\). Show that \(\{ f(x_n) \}\) has no subsequence that converges in \((0,1)\).

Let \((X,d_X)\) and \((Y,d_Y)\) be metric space and \(f \colon X \to Y\) be a one-to-one and onto continuous function. Suppose \(X\) is compact. Prove that the inverse \(f^{-1} \colon Y \to X\) is continuous.

Take the metric space of continuous functions \(C([0,1],{\mathbb{R}})\). Let \(k \colon [0,1] \times [0,1] \to {\mathbb{R}}\) be a continuous function. Given \(f \in C([0,1],{\mathbb{R}})\) define \[\varphi_f(x) := \int_0^1 k(x,y) f(y) ~dy .\] a) Show that \(T(f) := \varphi_f\) defines a function \(T \colon C([0,1],{\mathbb{R}}) \to C([0,1],{\mathbb{R}})\).

b) Show that \(T\) is continuous.

Let \((X,d)\) be a metric space.

a) If \(p \in X\), show that \(f \colon X \to {\mathbb{R}}\) defined by \(f(x) := d(x,p)\) is continuous.

b) Define a metric on \(X \times X\) as in part b, and show that \(g \colon X \times X \to {\mathbb{R}}\) defined by \(g(x,y) := d(x,y)\) is continuous.

c) Show that if \(K_1\) and \(K_2\) are compact subsets of \(X\), then there exists a \(p \in K_1\) and \(q \in K_2\) such that \(d(p,q)\) is minimal, that is, \(d(p,q) = \inf \{ (x,y) \colon x \in K_1, y \in K_2 \}\).

### Fixed point theorem and Picard’s theorem again

*Note: 1 lecture (optional, does not require )*

In this section we prove the fixed point theorem for contraction mappings. As an application we prove Picard’s theorem, which we proved without metric spaces in . The proof we present here is similar, but the proof goes a lot smoother with metric spaces and the fixed point theorem.

#### Fixed point theorem

Let \((X,d)\) and \((X',d')\) be metric spaces. \(f \colon X \to X'\) is said to be a *contraction* (or a contractive map) if it is a \(k\)-Lipschitz map for some \(k < 1\), i.e. if there exists a \(k < 1\) such that \[d'\bigl(f(x),f(y)\bigr) \leq k d(x,y) \ \ \ \ \text{for all } x,y \in X.\]

If \(f \colon X \to X\) is a map, \(x \in X\) is called a *fixed point* if \(f(x)=x\).

[Contraction mapping principle or Fixed point theorem] [thm:contr] Let \((X,d)\) be a nonempty complete metric space and \(f \colon X \to X\) a contraction. Then \(f\) has a unique fixed point.

The words *complete* and *contraction* are necessary. See .

Pick any \(x_0 \in X\). Define a sequence \(\{ x_n \}\) by \(x_{n+1} := f(x_n)\). \[d(x_{n+1},x_n) = d\bigl(f(x_n),f(x_{n-1})\bigr) \leq k d(x_n,x_{n-1}) \leq \cdots \leq k^n d(x_1,x_0) .\] Suppose \(m > n\), then \[\begin{split} d(x_m,x_n) & \leq \sum_{i=n}^{m-1} d(x_{i+1},x_i) \\ & \leq \sum_{i=n}^{m-1} k^i d(x_1,x_0) \\ & = k^n d(x_1,x_0) \sum_{i=0}^{m-n-1} k^i \\ & \leq k^n d(x_1,x_0) \sum_{i=0}^{\infty} k^i = k^n d(x_1,x_0) \frac{1}{1-k} . \end{split}\] In particular the sequence is Cauchy (why?). Since \(X\) is complete we let \(x := \lim\, x_n\), and we claim that \(x\) is our unique fixed point.

Fixed point? The function \(f\) is continuous as it is a contraction, so Lipschitz continuous. Hence \[f(x) = f( \lim \, x_n) = \lim\, f(x_n) = \lim\, x_{n+1} = x .\]

Unique? Let \(x\) and \(y\) both be fixed points. \[d(x,y) = d\bigl(f(x),f(y)\bigr) \leq k d(x,y) .\] As \(k < 1\) this means that \(d(x,y) = 0\) and hence \(x=y\). The theorem is proved.

The proof is constructive. Not only do we know a unique fixed point exists. We also know how to find it. We start with any point \(x_0 \in X\) and simply iterate \(f(x_0)\), \(f(f(x_0))\), \(f(f(f(x_0)))\), etc…In fact, you can even find how far away from the fixed point you are, see the exercises. The idea of the proof is therefore used in real world applications.

#### Picard’s theorem

Before we get to Picard, let us mention what metric space we will be applying the fixed point theorem to. We will use the metric space \(C([a,b],{\mathbb{R}})\) of . That is, \(C([a,b],{\mathbb{R}})\) is the space of continuous functions \(f \colon [a,b] \to {\mathbb{R}}\) with the metric \[d(f,g) = \sup_{x \in [a,b]} \left\lvert {f(x)-g(x)} \right\rvert .\] Convergence in this metric is convergence in uniform norm, or in other words, uniform convergence. Therefore, see , \(C([a,b],{\mathbb{R}})\) is a complete metric space.

Let us use the fixed point theorem to prove the classical Picard theorem on the existence and uniqueness of ordinary differential equations. Consider the equation \[\frac{dy}{dx} = F(x,y) .\] Given some \(x_0, y_0\) we are looking for a function \(y=f(x)\) such that \(f(x_0) = y_0\) and such that \[f'(x) = F\bigl(x,f(x)\bigr) .\] To avoid having to come up with many names we often simply write \(y' = F(x,y)\) and \(y(x)\) for the solution.

The simplest example is for example the equation \(y' = y\), \(y(0) = 1\). The solution is the exponential \(y(x) = e^x\). A somewhat more complicated example is \(y' = -2xy\), \(y(0) = 1\), whose solution is the Gaussian \(y(x) = e^{-x^2}\).

There are some subtle issues, for example how long does the solution exist. Look at the equation \(y' = y^2\), \(y(0)=1\). Then \(y(x) = \frac{1}{1-x}\) is a solution. While \(F\) is a reasonably “nice” function and in particular exists for all \(x\) and \(y\), the solution “blows up” at \(x=1\). For more examples related to Picard’s theorem see .

Let \(I, J \subset {\mathbb{R}}\) be compact intervals, let \(I_0\) and \(J_0\) be their interiors, and let \((x_0,y_0) \in I_0 \times J_0\). Suppose \(F \colon I \times J \to {\mathbb{R}}\) is continuous and Lipschitz in the second variable, that is, there exists an \(L \in {\mathbb{R}}\) such that \[\left\lvert {F(x,y) - F(x,z)} \right\rvert \leq L \left\lvert {y-z} \right\rvert \ \ \ \text{ for all $y,z \in J$, $x \in I$} .\] Then there exists an \(h > 0\) and a unique differentiable function \(f \colon [x_0 - h, x_0 + h] \to J \subset {\mathbb{R}}\), such that \[f'(x) = F\bigl(x,f(x)\bigr) \qquad \text{and} \qquad f(x_0) = y_0.\]

Without loss of generality assume \(x_0 =0\). As \(I \times J\) is compact and \(F(x,y)\) is continuous, it is bounded. So find an \(M > 0\), such that \(\left\lvert {F(x,y)} \right\rvert \leq M\) for all \((x,y) \in I\times J\). Pick \(\alpha > 0\) such that \([-\alpha,\alpha] \subset I\) and \([y_0-\alpha, y_0 + \alpha] \subset J\). Let \[h := \min \left\{ \alpha, \frac{\alpha}{M+L\alpha} \right\} .\] Note \([-h,h] \subset I\). Define the set \[Y := \{ f \in C([-h,h],{\mathbb{R}}) : f([-h,h]) \subset J \} . % [y_0-\alpha,y_0+\alpha] \} .\] That is, \(Y\) is the space of continuous functions on \([-h,h]\) with values in \(J\), in other words, exactly those functions where \(F\bigl(x,f(x)\bigr)\) makes sense. The metric used is the standard metric given above.

Show that \(Y \subset C([-h,h],{\mathbb{R}})\) is closed. Hint: \(J\) is closed.

The space \(C([-h,h],{\mathbb{R}})\) is complete, and a closed subset of a complete metric space is a complete metric space with the subspace metric, see . So \(Y\) with the subspace metric is complete.

Define a mapping \(T \colon Y \to C([-h,h],{\mathbb{R}})\) by \[T(f)(x) := y_0 + \int_0^x F\bigl(t,f(t)\bigr)~dt .\]

Show that if \(f \colon [-h,h] \to J\) is continuous then \(F\bigl(t,f(t)\bigr)\) is continuous on \([-h,h]\) as a function of \(t\). Use this to show that \(T\) is well defined and that \(T(f) \in C([-h,h],{\mathbb{R}})\).

Let \(f \in Y\) and \(\left\lvert {x} \right\rvert \leq h\). As \(F\) is bounded by \(M\) we have \[\begin{split} \left\lvert {T(f)(x) - y_0} \right\rvert &= \left\lvert {\int_0^x F\bigl(t,f(t)\bigr)~dt} \right\rvert \\ & \leq \left\lvert {x} \right\rvert M \leq hM \leq \frac{\alpha M}{M+ L\alpha} \leq \alpha . \end{split}\] So \(T(f)([-h,h]) \subset [y_0-\alpha,y_0+\alpha] \subset J\), and \(T(f) \in Y\). In other words, \(T(Y) \subset Y\). We thus consider \(T\) as a mapping of \(Y\) to \(Y\).

We claim \(T \colon Y \to Y\) is a contraction. First, for \(x \in [-h,h]\) and \(f,g \in Y\) we have \[\left\lvert {F\bigl(x,f(x)\bigr) - F\bigl(x,g(x)\bigr)} \right\rvert \leq L\left\lvert {f(x)- g(x)} \right\rvert \leq L \, d(f,g) .\] Therefore, \[\begin{split} \left\lvert {T(f)(x) - T(g)(x)} \right\rvert &= \left\lvert {\int_0^x F\bigl(t,f(t)\bigr) - F\bigl(t,g(t)\bigr)~dt} \right\rvert \\ & \leq \left\lvert {x} \right\rvert L \, d(f,g) \leq h L\, d(f,g) \leq \frac{L\alpha}{M+L\alpha} \, d(f,g) . \end{split}\] We chose \(M > 0\) and so \(\frac{L\alpha}{M+L\alpha} < 1\). The claim is proved by taking supremum over \(x \in [-h,h]\) of the left hand side above to obtain \(d\bigl(T(f),T(g)\bigr) \leq \frac{L\alpha}{M+L\alpha} \, d(f,g)\).

We apply the fixed point theorem () to find a unique \(f \in Y\) such that \(T(f) = f\), that is, \[%\label{equation:msinteqpicard} f(x) = y_0 + \int_0^x F\bigl(t,f(t)\bigr)~dt .\] By the fundamental theorem of calculus, \(T(f)\) is the unique differentiable function whose derivative is \(F\bigl(x,f(x)\bigr)\) and \(T(f)(0) = y_0\). Therefore \(f\) is the unique solution of \(f'(x) = F\bigl(x,f(x)\bigr)\) and \(f(0) = y_0\).

Prove that the statement “Without loss of generality assume \(x_0 = 0\)” is justified. That is, prove that if we know the theorem with \(x_0 = 0\), the theorem is true as stated.

## Exercises

For more exercises related to Picard’s theorem see .

Let \(F \colon {\mathbb{R}}\to {\mathbb{R}}\) be defined by \(F(x) := kx + b\) where \(0 < k < 1\), \(b \in {\mathbb{R}}\).

a) Show that \(F\) is a contraction.

b) Find the fixed point and show directly that it is unique.

Let \(f \colon [0,\nicefrac{1}{4}] \to [0,\nicefrac{1}{4}]\) be defined by \(f(x) := x^2\) is a contraction.

a) Show that \(f\) is a contraction, and find the best (smallest) \(k\) from the definition that works.

b) Find the fixed point and show directly that it is unique.

[exercise:nofixedpoint] a) Find an example of a contraction \(f \colon X \to X\) of non-complete metric space \(X\) with no fixed point. b) Find a 1-Lipschitz map \(f \colon X \to X\) of a complete metric space \(X\) with no fixed point.

Consider \(y' =y^2\), \(y(0)=1\). Use the iteration scheme from the proof of the contraction mapping principle. Start with \(f_0(x) = 1\). Find a few iterates (at least up to \(f_2\)). Prove that the pointwise limit of \(f_n\) is \(\frac{1}{1-x}\), that is for every \(x\) with \(\left\lvert {x} \right\rvert < h\) for some \(h > 0\), prove that \(\lim\limits_{n\to\infty}f_n(x) = \frac{1}{1-x}\).

Suppose \(f \colon X \to X\) is a contraction for \(k < 1\). Suppose you use the iteration procedure with \(x_{n+1} := f(x_n)\) as in the proof of the fixed point theorem. Suppose \(x\) is the fixed point of \(f\).

a) Show that \(d(x,x_n) \leq k^n d(x_1,x_0) \frac{1}{1-k}\) for all \(n \in {\mathbb{N}}\).

b) Suppose \(d(y_1,y_2) \leq 16\) for all \(y_1,y_2 \in X\), and \(k= \nicefrac{1}{2}\). Find an \(N\) such that starting at any point \(x_0 \in X\), \(d(x,x_n) \leq 2^{-16}\) for all \(n \geq N\).

Let \(f(x) := x-\frac{x^2-2}{2x}\). (You may recognize Newton’s method for \(\sqrt{2}\))

a) Prove \(f\bigl([1,\infty)\bigr) \subset [1,\infty)\).

b) Prove that \(f \colon [1,\infty) \to [1,\infty)\) is a contraction.

c) Apply the fixed point theorem to find an \(x \geq 1\) such that \(f(x) = x\), and show that \(x = \sqrt{2}\).

Suppose \(f \colon X \to X\) is a contraction, and \((X,d)\) is a metric space with the discrete metric, that is \(d(x,y) = 1\) whenever \(x \not= y\). Show that \(f\) is constant, that is, there exists a \(c \in X\) such that \(f(x) = c\) for all \(x \in X\).

- The term “modern” refers to late 19th century up to the present.↩
- For the fans of the TV show
*Futurama*, there is a movie theater in one episode called an \(\aleph_0\)-plex.↩ - An algebraist would say that \({\mathbb{Z}}\) is an ordered ring, or perhaps more precisely a commutative ordered ring.↩
- Uniqueness is up to isomorphism, but we wish to avoid excessive use of algebra. For us, it is simply enough to assume that a set of real numbers exists. See Rudin for the construction and more details.↩
- Named after the Swiss mathematician Jacob Bernoulli (1655 – 1705).↩
- The boundedness hypothesis is for simplicity, it can be dropped if we allow for the extended real numbers.↩
- use the notation \((x_n)\) to denote a sequence instead of \(\{ x_n \}\), which is what uses. Both are common.↩
- Named after the English physicist and mathematician Isaac Newton (1642 – 1726/7).↩
- Named after the Czech mathematician Bernhard Placidus Johann Nepomuk Bolzano (1781 – 1848), and the German mathematician Karl Theodor Wilhelm Weierstrass (1815 – 1897).↩
- Sometimes it is said that \(\{ x_n \}\)
*converges to infinity*.↩ - Named after the French mathematician Augustin-Louis Cauchy (1789–1857).↩
- The divergence of the harmonic series was known before the theory of series was made rigorous. In fact the proof we give is the earliest proof and was given by Nicole Oresme (1323?–1382).↩
- Demonstration of this fact is what made the Swiss mathematician Leonhard Paul Euler (1707 – 1783) famous.↩
- Named for the Italian mathematician Ernesto Cesàro (1859 – 1906).↩
- Named after the German mathematician Johannes Karl Thomae (1840 – 1921).↩
- Named after the German mathematician Rudolf Otto Sigismund Lipschitz (1832–1903).↩
- Named for the German mathematician Gottfried Wilhelm Leibniz (1646–1716).↩
- Named after the French mathematician Michel Rolle (1652–1719).↩
- Named for the English mathematician Brook Taylor (1685–1731). It was first found by the Scottish mathematician James Gregory (1638 – 1675). The statement we give was proved by Joseph-Louis Lagrange (1736 – 1813)↩
- Named after the German mathematician Georg Friedrich Bernhard Riemann (1826–1866).↩
- Named after the French mathematician Jean-Gaston Darboux (1842–1917).↩
- Such an \(h\) is said to be of
*bounded variation*.↩ - Compare this hypothesis to .↩
- Named for the Swiss mathematician Leonhard Paul Euler (1707 – 1783) and the Italian mathematician Lorenzo Mascheroni (1750 – 1800).↩
- Shortened from Latin:
*sinus cardinalis*↩ - The notation nor terminology is not completely standardized. The norm is also called the
*sup norm*or*infinity norm*, and in addition to \(\left\lVert {f} \right\rVert_u\) and \(\left\lVert {f} \right\rVert_S\) it is sometimes written as \(\left\lVert {f} \right\rVert_{\infty}\) or \(\left\lVert {f} \right\rVert_{\infty,S}\).↩ - Named for the French mathematician Charles Émile Picard (1856–1941).↩