Meeting time: 19:00 UTC on Saturdays and Sundays usually.^{†}
Meeting duration: 40 minutes.
Meeting link: susam.net/meet
Meeting log: click here
Reference book: Real Analysis by John M. Howie (2001)
Chapter notes: ch01.html
Club channel: #susam:matrix.org / #susam:libera.chat^{‡}
Mastodon: @susam@mastodon.social
Started: 27 Jul 2024
† There are some exceptions to this schedule occasionally. Join #susam:matrix.org or #susam:libera.chat or follow @susam@mastodon.social to receive schedule updates.
‡ You only need to join either the Matrix channel or the Libera channel, not both. Both channels are bridged together. If you are not an active IRC user, prefer joining the Matrix channel because it is more convenient for someone unfamiliar with IRC. For example, you can close your browser or client and your chat session will still stay alive on Matrix. You can connect back the next day and catch up with the messages. Doing that with IRC requires slightly more work such as setting up IRC bouncers etc.
The primary reference book for these meetings is Real Analysis written by the Scottish mathematician John Mackintosh Howie.
These meetings are hosted by Susam and attended by some members of
##math
channel of Libera IRC network as well as by some
members from Hacker
News.
You are welcome to join these meetings anytime. If you are concerned that the meetings may not make sense if you join when we are in the middle of a chapter, please free to talk to us about it in the group channel. I can recommend the next best time to begin joining the meetings. Usually, it would be when we begin reading a new section or chapter that is fairly self-contained and does not depend a lot on material we have read previously.
This is a hobby club with a focus on mathematics and computation. This club picks reading material about concepts and technologies that have been around for a long time and have an air of timelessness around them. See the blog post Reading Classic Computation Books for more details.
My name is Susam. This is my website. I run this club. I host the book club meetings. The last two series of book club meetings I hosted were about analytic number theory and Emacs. Some members of Libera IRC network and Hacker News participated in those meetings. This new series is going to be about real analysis.
We will discuss the content of the reference book. This includes topics like sequences and series, functions and continuity, differentiation and integration, Taylor's theorem, the Riemann integral, etc.
The first series of book club meeting I organised focussed on the book Introduction to Analytic Number Theory written by Tom M. Apostol. While the book excelled in rigour, some members, especially those without a strong mathematics background, found the constant alternation between theorems and proofs a bit too dry. Personally, I thoroughly enjoyed it. See Journey to Prime Number Theorem for a related post.
For the current series of meetings, I have chosen a more lightweight book that is easy to read. Howie's relaxed writing style in this book seems enjoyable. I do not know yet if this ease of reading comes at the cost of rigour. I might be able to assess that better as we make more progress through this book.
No, I have read the first few chapters of the book. This means that I am a few chapters ahead of someone who has just begun reading this book, so I can maintain a steady pace in the club discussions and also be able to tell if some topics or areas of confusion in an earlier chapter will be clarified in a later chapter.
See the meeting log which contains a rough plan for the next few meetings along with an archive of all previous meetings.
Not at all. It is up to you, really. If you would like to read the chapters in advance and come, that's great. But it is not necessary. We are going to discuss every page of the book in detail anyway.
Yes! Lurking is absolutely fine in our club meetings. In fact, most members of the club join in and stay silent throughout the meetings. Only a few members talk via audio/video or chat. This is considered absolutely normal in this club, so please do not hesitate to join our meetings!
Join the club channel at #susam:matrix.org or #susam:libera.chat to ask more questions.
Read on website | #mathematics | #real-analysis | #book | #meetup
]]>We get this absurd result because of the incorrect assumption that there exists an integer that is the largest of all the integers.
]]>Let us focus on \( \ln x \) again. Is it not peculiar? What does \( \ln \) stand for really? Logarithm natural? Sounds very unnatural.
Well, as a kid I learnt that \( \ln \) here stands for the Latin phrase "logarithmus naturalis". It is only recently that I bothered to verify if this expansion of \( \ln x \) that I learnt as a kid is really true. The most credible discussion of this that I could find online is this thread on Mathematics Stack Exchange: math.stackexchange.com/q/1694. The answer by Dan Velleman points us to page 277 of an 1875 book Lehrbuch der Mathematik by Anton Steinhauser. Quoting the relevant portion from the page:
Man pflegt nun, um Verwechslungen dieser beiden Systeme vorzubeugen, mit log.nat. a (gesprochen: logarithmus naturalis a) oder ln . a, oder am einfachsten mit la den natürlichen, mit log.brigg. a (gesprochen: Logarithmus briggus a) oder log.a, oder am einfachsten mit lg. a den gemeinen Logarithmus (von a) zu bezeichnen.
Translated to English, it says:
In order to prevent confusion between these two systems, people now use log.nat. a (pronounced: logarithmus naturalis a) or ln . a, or easiest with la the natural ones, with log.brigg. a (pronounced: logarithm briggus a) or log.a, or most simply with lg. a to denote the common logarithm (of a).
So it does look like what I learnt as a kid is correct and the earliest possible reference of this the Internet is able to find for us is the 1875 book quoted above.
]]>William Paul Thurston (October 30, 1946 – August 21, 2012) was an American mathematician. He was a pioneer in the field of low-dimensional topology and was awarded the Fields Medal in 1982 for his contributions to the study of 3-manifolds.
Thurston was a professor of mathematics at Princeton University, University of California, Davis, and Cornell University. He was also a director of the Mathematical Sciences Research Institute.
MathOverflow makes all answers posted to the website available under a Creative Commons license. In particular, all answers posted before 08 Apr 2011 (UTC) are available under the terms of the Creative Commons Attribution-ShareAlike 2.5 Generic (CC BY-SA 2.5) license. Thurston wrote the answer I am about to share on 30 Oct 2010. Due to the license terms, this post too is available under the terms of the same license.
Thurston posted his answer while replying to a MathOverflow question: What's a mathematician to do?. The question enquires about how an ordinary mathematician can contribute to mathematics. Thurston's answer from mathoverflow.net/a/44213 is reproduced below:
It's not mathematics that you need to contribute to. It's deeper than that: how might you contribute to humanity, and even deeper, to the well-being of the world, by pursuing mathematics? Such a question is not possible to answer in a purely intellectual way, because the effects of our actions go far beyond our understanding. We are deeply social and deeply instinctual animals, so much that our well-being depends on many things we do that are hard to explain in an intellectual way. That is why you do well to follow your heart and your passion. Bare reason is likely to lead you astray. None of us are smart and wise enough to figure it out intellectually.
The product of mathematics is clarity and understanding. Not theorems, by themselves. Is there, for example any real reason that even such famous results as Fermat's Last Theorem, or the Poincaré conjecture, really matter? Their real importance is not in their specific statements, but their role in challenging our understanding, presenting challenges that led to mathematical developments that increased our understanding.
The world does not suffer from an oversupply of clarity and understanding (to put it mildly). How and whether specific mathematics might lead to improving the world (whatever that means) is usually impossible to tease out, but mathematics collectively is extremely important.
I think of mathematics as having a large component of psychology, because of its strong dependence on human minds. Dehumanized mathematics would be more like computer code, which is very different. Mathematical ideas, even simple ideas, are often hard to transplant from mind to mind. There are many ideas in mathematics that may be hard to get, but are easy once you get them. Because of this, mathematical understanding does not expand in a monotone direction. Our understanding frequently deteriorates as well. There are several obvious mechanisms of decay. The experts in a subject retire and die, or simply move on to other subjects and forget. Mathematics is commonly explained and recorded in symbolic and concrete forms that are easy to communicate, rather than in conceptual forms that are easy to understand once communicated. Translation in the direction conceptual -> concrete and symbolic is much easier than translation in the reverse direction, and symbolic forms often replaces the conceptual forms of understanding. And mathematical conventions and taken-for-granted knowledge change, so older texts may become hard to understand.
In short, mathematics only exists in a living community of mathematicians that spreads understanding and breaths life into ideas both old and new. The real satisfaction from mathematics is in learning from others and sharing with others. All of us have clear understanding of a few things and murky concepts of many more. There is no way to run out of ideas in need of clarification. The question of who is the first person to ever set foot on some square meter of land is really secondary. Revolutionary change does matter, but revolutions are few, and they are not self-sustaining --- they depend very heavily on the community of mathematicians.
In the comments to the answer, one of the commenters was Suresh Venkatasubramanian who was a professor in the School of Computing at the University of Utah back then. He is now a professor of Computer Science and Data Science at Brown University. In his comment, Suresh proposed that this answer be called Thurston's Paean. Here is his complete comment:
This seems like an ideal counterpoint to Hardy's Lament. I'm calling it Thurston's Paean :). Seems poignant now that he has passed.
Thurston's answer does appear to be a perfect complement to Hardy's lament in the 1940 essay A Mathematician's Apology. While Hardy's lament is remarkably beautiful and introspective, it may also feel a little depressing at places. Thurston's post on the other hand is full of hope and purpose that goes beyond the actual work of doing mathematics. Indeed Thurston's Paean is a befitting title for his answer.
]]>One of the many techniques for solving ordinary differential equations involves using an integrating factor. An integrating factor is a function that a differential equation is multiplied by to simplify it and make it integrable. It almost appears to work like magic!
Let us first see how the integrating factor method works. In this post, we will work with linear first-order ordinary differential equations of type \[ \frac{dy}{dx} + y P(x) = Q(x) \] to discuss, reason about, and illustrate this method. We will also often use the Leibniz's notation \( dy/dx \) and the Lagrange's notation \( y'(x) \) or simply \( y' \) interchangeably as is typical in calculus. They all mean the same thing: the derivative of the function \( y \) with respect to \( x. \) Thus the above differential equation may also be written as \[ y' + y P(x) = Q(x). \] Given a differential equation of this form, we first find an integrating factor \( M(x) \) using the formula \[ M(x) = e^{\int P(x) \, dx}. \] Then we multiply both sides of the differential equation by this integrating factor. Now remarkably, the left-hand side (LHS) reduces to a single term consisting only of a derivative. As a result, we can get rid of that derivative by integrating both sides of the equation and we then proceed to obtain a solution.
Here is an example that demonstrates the method of using an integrating factor. Let us say we want to solve the differential equation \[ y' + y \left( \frac{x + 1}{x} \right) = \frac{1}{x}. \] Indeed this is in the form \( y' + y P(x) = Q(x) \) with \( P(x) = (x + 1)/x \) and \( Q(x) = 1/x. \) We first obtain the integrating factor \[ M(x) = e^{\int P(x) \, dx} = e^{\int (x + 1)/x \, dx} = e^{\int (1 + 1/x) \, dx} = e^{x + \ln x} = x e^x. \] Now we multiply both sides of the differential equation by this integrating factor and get \[ y' x e^x + y (x + 1) e^x = e^x. \] The LHS can now be simplified to \( \frac{d}{dx} (y x e^x). \) This can be verified using the product rule for derivatives. This simplification of the LHS is the remarkable feature of this method. Therefore the above equation can be written as \[ \frac{d}{dx} (y x e^x) = e^x. \] Note that the expression on the LHS is a product of the function \( y \) and the integrating factor \( x e^x. \) We will discuss this observation in more detail a little later. Let us first complete solving this differential equation. Since the LHS is now a single term that consists of a derivative, obtaining a solution now simply involves integrating both sides with respect to \( x. \) Integrating both sides we get \[ y x e^x = e^x + C \] where \( C \) is the constant of integration. Finally, we divide both sides by the integrating factor \( x e^x \) to get \[ y = \frac{1}{x} + \frac{C}{x e^x}. \] We have now obtained a solution for the differential equation. If we review the steps above, we will find that after multiplying both sides of the given differential equation by the integrating factor, the differential equation becomes significantly simpler and integrable. In fact, after multiplying both sides of the given differential equation by the integrating factor, the LHS always becomes the derivative of the product of the function \( y \) and the integrating factor. We will now see why this is so.
Consider once again the linear first-order differential equation \begin{equation} \label{if-eq-diff} y' + yP(x) = Q(x). \end{equation} We first find the integrating factor \begin{equation} \label{if-eq-integrating-factor} M(x) = e^{\int P(x)\, dx}. \end{equation} The integrating factor obtained like this satisfies an interesting relationship: \begin{equation} \label{if-eq-property} M'(x) = M(x) P(x). \end{equation} We can prove this relationship easily by differentiating both sides of \eqref{if-eq-integrating-factor} as follows: \[ M'(x) = \frac{d}{dx} \left( e^{\int P(x)\, dx} \right) = e^{\int P(x)\, dx} \frac{d}{dx} \left( \int P(x)\, dx \right) = M(x) P(x). \] Note that we use the chain rule to work out the derivative above. This beautiful result is due to how the derivative of the exponential function works. When we apply the chain rule to obtain the derivative of \( e^{f(x)} \) we get \[ \frac{d}{dx} e^{f(x)} = e^{f(x)} f'(x). \] This nice property of the exponential function leads to the interesting relationship in \eqref{if-eq-property}.
Now let us multiply both sides of the differential equation \eqref{if-eq-diff} by the integrating factor \( M(x) \) By doing so, we get \[ y' M(x) + y P(x) M(x) = Q(x) M(x). \] But from \eqref{if-eq-property} we know that \( P(x) M(x) = M'(x), \) so the above equation can be written as \[ y' M(x) + y M'(x) = Q(x) M(x). \] Look what we have got on the LHS! We have the expansion of \( \frac{d}{dx}(yM(x)) \) on the LHS. By product rule of differentiation, we have \( \frac{d}{dx}(yM(x)) = y' M(x) + y M'(x). \) Therefore the above equation can be written as \[ \frac{d}{dx}(yM(x)) = Q(x) M(x). \] The "magic" has occurred here! Multiplying both sides of the differential equation by the integrating factor has led us to an equation that has got a single derivative only on the LHS. As a result, finding the solution is now a simple matter of integrating both sides, i.e., \[ y M(x) = \int Q(x) M(x) \, dx. \] Thus \[ y = \frac{1}{M(x)} \int Q(x) M(x) \, dx. \] Note that the result of indefinite integral on the RHS will contain the constant of integration, which we will denote as \( C, \) so the final solution looks like \begin{equation} \label{if-eq-general-solution} y = \frac{1}{M(x)} \int Q(x) M(x) \, dx + \frac{C}{M(x)}. \end{equation}
Let us illustrate the method and its magic with a very simple differential equation: \[ y' + \frac{y}{x} = x. \] First we note that this equation is in the form \( y' + yP(x) = Q(x) \) with \( P(x) = 1/x \) and \( Q(x) = x. \) We then find the integrating factor \[ M(x) = e^{\int P(x) \, dx} = e^{\int \frac{1}{x} \, dx} = e^{\ln x} = x. \] Then we multiply both sides of the differential equation by the integrating factor to get \[ y'x + y = x^2. \] Now indeed the LHS can be written down as a single derivative as shown below: \[ \frac{d}{dx} yx = x^2. \] Note that the LHS is the derivative of the product of \( y \) and the integrating factor \( x. \) This is exactly what we discussed in the previous section. We integrate both sides of the above equation to get \[ yx = \frac{x^3}{3} + C. \] Finally we divide both sides by the integrating factor \( x \) to get \[ y = \frac{x^2}{3} + \frac{C}{x}. \] We have arrived at the solution \( y(x) \) for the differential equation.
In this post, we used very simple and convenient differential equations that led to nice closed-form solutions. In practice, differential equations can be quite complicated and may not always lead to closed-form solutions. In such cases, we leave the result in the form of an expression that contains an unsolved integral. Such solutions may resemble the form shown in \eqref{if-eq-general-solution}.
The method of using integrating factors to solve differential equations can also be extended to linear higher-order differential equations. That is something we did not discuss in this post. However, I hope that the intuition gained from understanding how and why this method works for linear first-order differential equations will be useful while studying such extensions of this method.
]]>One of the things about the book that caught my interest from the very beginning was its front cover. It has a peculiarly drawn grid of white boxes and red empty regions that looks quite interesting. Here is the grid from the front cover of the book:
Can we come up with a simple and elegant rule that defines this grid? Here is one I could come up with:
We define \( \gcd(x, y) \) to be a nonnegative common divisor of \( x \) and \( y \) such that every common divisor of \( x \) and \( y \) also divides \( \gcd(x, y). \) Let us now see if we can explain some of the interesting properties of this grid using the above rule:
When \( x = 0 \) and \( y \ne 1, \) we get \( \gcd(x, y) = \lvert y \rvert \ne 1, \) so the entire column at \( x = 0 \) has boxes except at \( (0, 1). \) Similarly, the entire row at \( y = 0 \) has boxes except at \( (1, 0). \)
The cell \( (0, 0) \) has a box because \( \gcd(0, 0) \ne 1. \) In fact, \( \gcd(0, 0) = 0. \) This follows from the definition of the \( \gcd \) function. We will discuss this in more detail later in this post.
Every diagonal cell \( (x, x) \) has a box except at \( (1, 1) \) because \( \gcd(x, x) = \lvert x \rvert \) for all integers \( x. \)
The grid is symmetric about the diagonal cells \( (x, x) \) because \( \gcd(x, y) = \gcd(y, x). \)
A column at \( x \) has exactly one cell below the diagonal if and only if \( x \) is prime. For example, check the column for \( x = 5. \) It has exactly one cell below the diagonal. We know that \( 5 \) is prime. Now check the column for \( x = 6. \) It has four cells below the diagonal. We know that \( 6 \) is not prime.
Let us now elaborate the second point in the list above. If \( \gcd(0, 0) \) is \( 0, \) then \( 0 \) must divide \( 0. \) Does \( 0 \) really divide \( 0? \) Isn't \( 0/0 \) undefined? Yes, even though \( 0/0 \) is undefined, \( 0 \) divides \( 0. \) We say an integer \( d \) divides an integer \( n \) when \( n = cd \) for some integer \( c. \) We have \( 0 = 0 \cdot 0, \) so indeed \( 0 \) divides \( 0. \)
We have shown that \( 0 \) divides \( 0 \) but we have not shown yet that \( \gcd(0, 0) = 0. \) Is \( \gcd(0, 0) \) really \( 0? \) Every integer divides \( 0, \) e.g., \( 1 \) divdes \( 0, \) \( 2 \) divides \( 0, \) \( 3 \) divides \( 0, \) etc. There does not seem to be a greatest common divisor of \( 0 \) and \( 0. \) Shouldn't \( \gcd(0, 0) \) be called either infinity or undefined? No, we need to look at the definition of \( \gcd \) introduced earlier. As per the definition, every common divisor of integers \( x \) and \( y \) must also divide \( \gcd(x, y). \) With this requirement in mind, we see that \( \gcd(0, 0) \) must be \( 0. \) This definition also makes \( \gcd(n, 0) = \gcd(0, n) = \lvert n \rvert \) for all integers \( n. \) Further, this definition makes Bézout's identity hold for all integers. Bézout's identity states that there exists integers \( m \) and \( n \) such that \( mx + ny = \gcd(x, y). \) Indeed if we have \( \gcd(0, 0) = 0, \) we get \( 0 \cdot 0 + 0 \cdot 0 = 0 = \gcd(0, 0). \)
That's all I wanted to share about the front cover of the book. While the front cover is quite interesting, the content of the book is even more fascinating. I found chapters 12 and 13 of the book to be the most interesting. In chapter 12, the book teaches how to prove that the Riemann zeta function \( \zeta(s) \) vanishes at every negative even integer \( s. \) Through several contour integrals and clever use of Cauchy's residue theorem, it shows in the end that \( \zeta(-2n) = 0 \) for \( n = 1, 2, 3, \dots. \) In chapter 13, the book shows us how to obtain zero-free regions where \( \zeta(s) \) does not vanish. The book exposes various subtle nuances of the zeta function with great rigour and thoroughness. Results like \( \zeta(-1) = 1/12 \) that once felt mysterious look crystal clear and obvious after working through this book. I strongly recommend this book to anyone who wants to learn analytic number theory.
]]>We have been reading the book Introduction to Analytic Number Theory by Apostol (1976) since March 2021. It has been going consistently since then and the previous few posts on this blog provide an account of how this journey has been so far. After about seven months of reading this book together, we are having our final meeting for this book today. This is going to be the 120th meeting of our book discussion group. The meeting notes from all previous reading sessions are archived at IANT Notes. We will discuss the final two pages of this book today and complete reading this book.
In the meeting today, we will look at some applications of the recursion formula related to partition functions that we learnt earlier. Here is an excerpt from the book that shows a specific example that demonstrates the richness and beauty of concepts one can discover while studying analytic number theory:
Equation (24) becomes \[ np(n) = \sum_{k=1}^n \sigma(k) p(n - k). \] a remarkable relation connecting a function of multiplicative number theory with one of additive number theory.
Now what equation (24) contains is not important for this post. Of course, you can refer to the book if you really want to know what equation (24) is. We learnt to prove that equation in the penultimate meeting for this subject yesterday. In this post, I will emphasise how indeed this equation is remarkable.
The divisor sum function \( \sigma(n) \) represents the sum of all positive divisors of \( n. \) Here are some examples: \begin{align*} \sigma(1) &= 1, \\ \sigma(2) &= 1 + 2 = 3, \\ \sigma(3) &= 1 + 3 = 4, \\ \sigma(4) &= 1 + 2 + 4 = 7, \\ \sigma(5) &= 1 + 5 = 6. \end{align*} We have spent a good amount of time with this function in the initial chapters of the book. However, for the purpose of this blog post, the definition and the examples above are good enough.
The \( p(n) \) function is the unrestricted partition function. It represents the number of ways \( n \) can be written as a sum of positive integers \( \le n. \) Further, we let \( p(0) = 1. \) Here are some examples: \begin{align*} p(1) &= 1, \\ p(2) &= 2, \\ p(3) &= 3, \\ p(4) &= 4, \\ p(5) &= 7. \end{align*} Let me illustration the last value. The integer \( 5 \) can be represented as a sum of positive integers \( \le 5 \) in 7 different ways. They are: \( 5, \) \( 4 + 1, \) \( 3 + 2, \) \( 3 + 1 + 1, \) \( 2 + 2 + 1, \) \( 2 + 1 + 1 + 1, \) and \( 1 + 1 + 1 + 1 + 1. \) Thus \( p(n) = 5. \)
The divisor sum function comes from multiplicative number theory. The partition function comes from additive number theory. Yet these two very different things get linked together in the formula mentioned in the excerpt included above. Here is the formula once again: \[ np(n) = \sum_{k=1}^n \sigma(k) p(n - k). \] How beautiful! How nicely the divisor sum function and the unrestricted partition function appear together elegantly in a single equation! Further, this equation provides a recursion formula for the partition function. Here is an illustration of this equation with \( n = 5 \): \[ 5 \cdot p(5) = 5 \cdot 7 = 35. \] \begin{align*} \sum_{k=1}^5 \sigma(k) p(5 - k) &= \sigma(1) p(4) + \sigma(2) p(3) + \sigma(3) p(2) + \sigma(4) p(1) + \sigma(5) p(0) \\ &= (1)(5) + (3)(3) + (4)(2) + (7)(1) + (6)(1) \\ &= 5 + 9 + 8 + 7 + 6 \\ &= 35. \end{align*} We will go through this topic once more in the meeting today, so if you are interested to see this formula worked out in a step-by-step manner, do join our final meeting for this book.
The final meeting is coming up at 17:00 UTC today. Visit the analytic number theory page to get the meeting link. This is not going to be the final meeting for our overall book discussion group though. This is going to be the finally meeting for only the analytic number theory book. We will have more meetings for another book after a short break.
The meeting today is going to be a lightweight session. The last two pages that we will discuss today contain some examples of recursion formulas and some commentary about Ramanujan's partition identities. Most of it should make sense even to those who have not been part of our meetings earlier, so everyone is welcome to join this meeting today, even if only to lurk. You can also join our group by joining our IRC channel where we will publish updates about future meetings. Our channel details are available in the main page here.
A big thank you to the Hacker News community and the Libera IRC mathematics and algorithms communities who showed interest in these meetings, joined the meetings, and made this series of meetings successful.
]]>After 114 meetings and 75 hours of studying together, our analytic number theory book discussion group has finally reached the final chapter of the book Introduction to Analytic Number Theory by Apostol (1976). We have less than 18 pages to read in order to complete reading this book. Considering that we meet 3-4 times in a week and we discuss about 2-3 pages in every meeting, it appears that we would be able to complete reading this book in another 2 weeks.
Reading this book has been quite a journey! The previous three posts on this blog provide an account of how this journey has been. It has been fun, of course. The best part of hosting a book discussion group like this has been the number of extremely smart people I got an opportunity to meet and interact with. The insights and comments on the study material that others shared during the meetings were very helpful.
The meeting log shows that our meetings started really small with only 4 participants in the first meeting in March 2021 and then it gradually grew to about 10-12 regular members within a month. Then a few months later, the number of participants began dwindling a little. This happened because some members of the group had to drop out as they got busy with other personal or professional engagements. However, six months later, we still have about 4-5 regular participants meeting consistently. I think it is pretty good that we have made it this far.
The final chapter on integer partitions is very unlike all the previous 12 chapters. While the previous chapters dealt with multiplicative number theory, this final chapter deals with additive number theory. For example, the first theorem talks about an interesting property of unrestricted partitions. We study the number of ways a positive integer can be expressed as a sum of positive integers. The number of summands is unrestricted, repetition of summands is allowed, and the order of the summands is not taken into account. For example, the number 3 has 3 partitions: 3, 2 + 1, and 1 + 1 + 1. Similarly, the number 4 has 5 partitions: 4, 3 + 1, 2 + 2, 2 + 1 + 1, and 1 + 1 + 1 + 1.
I have always wanted to learn about partitions more deeply, so I am quite happy that this book ends with a chapter on partitions. The subject of partitions is rich with very interesting results obtained by various accomplished mathematicians. In the book, the first theorem about partitions is a very simple one that follows from the geometric representation of partitions. Let us see an illustration first.
How many partitions of 6 are there? There are 11 partitions of 6. They are 6, 5 + 1, 4 + 2, 4 + 1 + 1, 3 + 3, 3 + 2 + 1, 3 + 1 + 1 + 1, 2 + 2 + 2, 2 + 2 + 1 + 1, 2 + 1 + 1 + 1 + 1, and 1 + 1 + 1 + 1 + 1 + 1. Now how many of these partitions are made up of 5 parts? Each summand is called a part. The answer is 2. There are 2 partitions of 6 that are made up of 5 parts. They are 3 + 1 + 1 + 1 and 2 + 2 + 1 + 1. Let us represent both these partitions as arrangements of lattice points. Here is the representation of the partition 3 + 1 + 1 + 1:
• • •
•
•
•
Now if we read this arrangement from left-to-right, column-by-column, we get another partition of 6, i.e., 4 + 1 + 1. Note that the number of parts in 3 + 1 + 1 + 1 (i.e., 4) appears as the largest part in 4 + 1 + 1. Similarly, the number of parts in 4 + 1 + 1 (i.e., 3) appears as the largest part in 3 + 1 + 1 + 1. Let us see one more example of this relationship. Here is the geometric representation of 2 + 2 + 1 + 1:
• •
• •
•
•
Once again, reading this representation from left-to-right, we get 4 + 2, another partition of 6. Once again, we can see that the number of partitions in 2 + 2 + 1 + 1 (i.e., 4) appears as the largest part in 4 + 2, and vice versa. These observations lead to the first theorem in the chapter on partitions:
Theorem 14.1 The number ofpartitions of \( n \) into \( m \) parts is equal to the number of partitions of \( n \) into parts, the largest of which is \( m. \)
That was a brief introduction to the chapter on partitions. In the next two or so weeks, we will dive deeper into the theory of partitions.
If this blog post was fun for you, consider joining our next meeting. Our next meeting is on Tue, 21 Sep 2021 at 17:00 UTC. Since we are at the beginning of a new chapter, it is a good time for new participants to join us. It is also a good time for members who have been away for a while to join us back. Since this chapter does not depend much on the previous chapters, new participants should be able to join our reading sessions for this chapter and follow along easily without too much effort.
To join our discussions, see our channel details in the main page here. To get the meeting link for the next meeting, visit the analytic number theory book page.
It is worth mentioning here that lurking is absolutely fine in our meetings. In fact, most participants of our meetings join in and stay silent throughout the meeting. Only a few members talk via audio/video or chat. This is considered absolutely normal in our meetings, so please do not hesitate to join our meetings!
]]>The book I had chosen for our discussions was Introduction to Analytic Number Theory by Apostol (1976). I have been hosting 40-minute meetings for about 3-4 days every week since March 2021. We discuss a couple of pages of the book in every meeting. Most participants in this meeting are from Hacker News and Libera IRC network. For a long time, I was eager to learn the proof of the prime number theorem. For those unfamiliar with the theorem, I will describe it briefly in further sections. Let me first answer the question I asked in the previous paragraph.
So how long does it take to start with no knolwedge of analytic number theory and teach ourselves the analytic proof of the prime number theorem? Turns out, it takes 72 hours! It took our group 72 hours spread across 110 meetings over 6 months to be able to understand the proof. It is worth noting here that most of us in this group have full-time jobs and other personal obligations! We were all doing this for fun, for the joy of learning!
Now I must mention that the 72 hours noted above is only the time spent together in reading the book and working through the theorems and proofs. It does not include the personal time spent in solving problems, reading some sections again, taking notes, etc. All of that was done in our personal time. We did discuss the solutions to some of the very interesting problems in our meetings just to take a break from the theorem-and-proof style of reading but most of these 72 hours of meetings focussed on working through the theorems and proofs in the book.
It may be possible to achieve this milestone in lesser number of hours, perhaps by reading the book alone which for some folks might be faster than studying in a group, or perhaps by skipping some chapters for topics that look very familiar. In our discussions, however, we did not skip any chapter. There were in fact a few chapters we could have skipped. All members of these meetings were very familiar with divisibility, greatest common divisor, the fundamental theorem of arithmetic, etc. discussed in Chapter 1. Most of us were also very familiar with the concepts discussed in Chapter 5 such as congruences, residue classes, the Euler-Fermat theorem, the Chinese remainder theorem, etc. Despite being familiar with these concepts, we decided not to skip any chapter for the sake of completeness of our coverage of the material. In fact, we read every single line of the book and deliberated over every single concept discussed in the book. With this detailed and tedious approach to reading the book, it took us 72 hours to read about 290 pages and learn the analytic proof of the prime number theorem in Chapter 13.
The prime number theorem is a very curious fact about the distribution of prime numbers that Gauss noticed in the year 1792 when he was about 15 years old. He noticed that the occurrence of primes become rarer and rarer as we expand our search for them to larger and larger integers. For example, there are 4 primes between 1 and 10, i.e., 40% of the numbers between 1 and 10 are primes! But there are only 25 primes between 1 and 100, i.e., only 25% of the numbers between 1 and 100 are primes. If we go up to 1000, we notice that there are only 168 primes between 1 and 1000, i.e., only 16.8% of the numbers between 1 and 1000 are primes. Formally, we denote these facts with the mathematical notation \( \pi(x) \) that denotes the prime counting function. We say \( \pi(10) = 4, \) \( \pi(100) = 25, \) \( \pi(1000) = 168, \) and so on. Note that we allow \( x \) to be a real number, so while \( \pi(10) = 4, \) we have \( \pi(10.3) = 4 \) as well. One of the reasons we let \( x \) be a real number in the definition of \( \pi(x) \) is because it makes various problems we come across during the study of this function more convenient to work on using real analysis.
We observe that the "density" of primes continue to fall as we make \( x \) larger and larger. In formal notation, we see that the ratio \( \pi(x) / x \) is \( 0.4 \) when \( x = 10. \) This ratio falls to \( 0.25 \) when \( x = 100. \) It falls further to \( 0.168 \) when \( x = 1000, \) and so on. Can we predict by how much this "density" falls? The answer is, yes, and that leads us to the prime number theorem. The prime number theorem states that \( \pi(x) / x \) is asymptotic to \( 1 / \log x \) as \( x \) approaches infinity, i.e., \[ \frac{\pi(x)}{x} \sim \frac{1}{\log x} \text{ as } x \to \infty. \] For those unfamiliar with the notation of asymptotic equality, here is another equivalent way to state the above relationship, \[ \lim_{x \to \infty} \frac{\pi(x) / x}{1 / \log x} = 1. \] We could also write this as \[ \lim_{x \to \infty} \frac{\pi(x)}{x / \log x} = 1 \] or \[ \pi(x) \sim \frac{x}{\log x} \text{ as } x \to \infty. \] Let us see how well this formula works as an estimate for the density of primes for small values of \( x. \)
\( x \) | \( \pi(x) \) | \( x / \log x \) |
---|---|---|
10 | 4 | 4.3 |
100 | 25 | 21.7 |
1000 | 168 | 144.8 |
10000 | 1229 | 1085.7 |
100000 | 9592 | 8685.9 |
Not bad! In fact, the last two columns begin to agree more and more as \( x \) becomes larger and larger.
The analytic proof of the prime number theorem was achieved with an intricate chain of equivalences and implications between various theorems. The book consumes 13 chapters and 290 pages before completing the proof of the prime number theorem. Each page is also quite dense with information. The amount of commentary or illustrations is very little in the book. Most of the book keeps alternating between theorem statements and proofs. Occasionally, for especially long chapters with an intricate sequence of proofs, Apostol provides a plan of the proof in the introductions to such chapters. It is quite hard to summarise a large and dense volume of work like this in a blog post but I will make an attempt to paint a very high-level picture of some of the key concepts that are involved in the proof.
Everything from Chapters 1 to 3 is about building basic concepts and tools we will use later to work on the problem of the prime number theorem. These concepts and tools were very interesting on their own. They involved divisibility, various number-theoretic functions, Dirichlet products, the big oh notation, etc. Chapter 4 was the first chapter where we engaged ourselves with the prime number theorem. This chapter taught us several other formulas that were logically equivalent to the prime number theorem. One equivalence that would play a big role later was the equivalence between the prime number theorem \[ \lim_{x \to \infty} \frac{\pi(x) \log x}{x} = 1 \] and the following form: \[ \lim_{x \to \infty} \frac{\psi(x)}{x} = 1. \] If we could prove one, the validity of the other would be established automatically. The notation \( \psi(x) \) denotes the Chebyshev function which in turn is defined in terms of the Mangoldt function \( \Lambda(n) \) as \( \psi(x) = \sum_{n \le x} \Lambda(n). \) Note that the formula above can also be stated using the asymptotic equality notation as follows: \[ \psi(x) \sim x \text{ as } x \to \infty. \] There were several other equivalent forms too shown in Chapter 4. The fact that all these various forms were equivalent to each other was rigorously proved in the chapter. Thus proving any one of the equivalent forms would be sufficient to prove the prime number theorem. But in Chapter 4, we did not know how to prove any of the equivalent forms. We could only prove the equivalence of the various formulas, not the formulas themselves. We only learnt that if any of the equivalent forms is true, so is the prime number theorem. Similarly, if any of the equivalent forms is false, so is the prime number theorem. We would visit the prime number theorem again in Chapter 13 which would complete the proof of the prime number theorem by showing that the equivalent form mentioned above is indeed true.
Chapters 5 to 10 introduced more concepts involving congruences, finite abelian groups, their characters, Dirichlet characters, Dirichlet's theorem on primes in arithmetic progressions, Gauss sums, quadratic residues, primitive roots, etc. Some of these concepts would turn out to be very important in proving the prime number theroem but most of them probably are not too important if understanding the proof of the prime number theorem is the only goal. Regardless, all of these chapters were very interesting.
It was in Chapters 11 and 12 that we felt that we were getting closer and closer to the proof of the prime number theorem. Chapter 11 began a detailed and rigorous study of convergence and divergence of Dirichlet series. The Riemann zeta function is a specific type of Dirichlet series. Chapter 12 introduced analytic continuation of the Riemann zeta function. We could then show interesting results like \( \zeta(0) = -1/2 \) and \( \zeta(-1) = -1/12 \) using the analytic continuation of the zeta function. This chapter also showed us why all trivial zeroes of \( \zeta(s) \) must lie at negative even integers.
One thing I realised during the study of this book is how frequently we use concepts, operations, functions, and theorems named after Dirichlet. It was impossible to get through a meeting without having uttered "Dirichlet" at least a dozen times!
Finally, Chapter 13 showed us how to prove the prime number theorem. The plan of the proof was laid out in the first section. Our goal in this chapter is to prove that \( \psi(x) \sim x \) as \( x \to \infty. \) This is equivalent to the prime number theorem, so proving this amounts to proving the prime number theorem too.
Next we learn that the asymptotic relation \( \psi_1(x) \sim x^2 / 2 \) as \( x \to \infty \) implies the previous asymptotic relationship. Here \( \psi_1(x) \) is defined as \( \psi_1(x) = \int_1^x \psi(t) \, dt. \) This implication is proved quite easily in one and a half pages. But we still need to show that the asymptotic relation \( \psi_1(x) \sim x^2 / 2 \) as \( x \to \infty \) indeed holds good. Proving this takes a lot of work. To prove this asymptotic relation we first learn to arrive at the following equation involving a contour integral: \[ \frac{\psi_1(x)}{x^2} - \frac{1}{2} \left( 1 - \frac{1}{x} \right)^2 = \frac{1}{2\pi i} \int_{c - \infty i}^{c + \infty i} \frac{x^{s - 1}}{s(s + 1)} \left( -\frac{\zeta'(s)}{\zeta(s)} - \frac{1}{s - 1} \right) \, ds \] for \( c > 1. \) The equation above looks quite complex initially but each part of it becomes friendly as we learn to derive it and then work on each part of it while working out further proofs. Now if we could somehow show that the integral on the right hand side of the above equation approaches 0 as \( x \to \infty, \) that would end up proving the asymptotic relation involving \( \psi_1(x) \) and thus end up proving the prime number theorem by equivalence. However, proving that this integral indeed becomes 0 as \( x \to \infty \) requires a careful study of \( \zeta(s)/\zeta'(s) \) in the vicinity of the line \( \operatorname{Re}(s) = 1. \) This is the topic that most of the chapter deals with.
This plan of the proof looked quite convoluted initially but Apostol has done a great job in this chapter to first walk us through this plan and then prove each fact that we need to make the proof work in a detailed and rigorous manner. When we reached the end of the proof, one of our regular members remarked, "Now the proof does not look so complex!"
Would the elementary proof of the prime number theory have been easier? I don't know. I have not studied the elementary proof. But Apostol does say this at the beginning of Chapter 13,
The analytic proof is shorter than the elementary proof sketched in Chapter 4 and its principal ideas are easier to comprehend.
Learning the analytic proof itself was quite a long journey that required dedication and consistency in our studies over a period of 6 months. If we trust the above excerpt from the book, then I think it is fair to assume that the elementary proof is even more formidable.
That was an account of our journey through an analytic number theory book from its first chapter up to the analytic proof of the prime number theorem. We have not completed reading the entire book though. We still have about another 30 pages to go through. In the remaining study of this book, we will learn more about zero-free regions for \( \zeta(s), \) the application of the prime number theorem to the divisor function, and the Euler totient function. The next and the final chapter too has a lot to offer such as integer partition, Euler's pentagonal-number theorem, and the partition identities of Ramanujan. I am pretty hopeful that we will be complete reading this book in another few weeks of meetings.
]]>In this blog post, I will talk about my personal experience hosting these meetings and my personal journey about reading this book. It is worth keeping in mind then that what I am about to write below may not have any resemblance with the experience of other participants of these meetings.
As far as I know, everyone who joins our meetings are involved in computer programming in one form or another. A few of them have very strong background in mathematics. I host these meetings everyday and discuss a few sections of the book in detail. I show how to work through the proofs, explain some of the steps, etc. Sometimes I get stuck in some step that I find too unobvious. Sometimes the steps are obvious but my brain is too slow to understand why the steps work. But these tiny glitches have not been a problem so far, thanks to all the members who join these meetings on a daily basis and contribute their explanations of the proofs.
I believe the group members are the best part of these discussions. Thanks to the insights and explanation of the reading material shared by all these members, I am fairly confident that we are able to take a close look at every proof and convince ourselves that every step of the proofs work.
The first web meeting to discuss the chosen analytic number theory book occurred on 5 Mar 2021. See the blog post Reading Classic Computation Books to read about the early days of our group and how it was formed. Back then, I knew little to nothing about analytic number theory. Although I was familiar with some of the elementary concepts like divisibility, Euler's totient function, modular arithmetic, calculus, and related theorems, chapter 2 of the book itself proved to be a significant challenge for me. In the second chapter, it became clear to me that we will be building new levels of mathematical abstractions, use these abstractions to build yet another layer of abstractions, and so on. The chapter began with a description of the Möbius function, a very neat and interesting function that I was previously unaware of. That was fun! But soon, this chapter began adding new layers of abstractions such as Dirichlet product, Dirichlet inverse, generalised convolution, etc. I could almost feel my brain stretching and growing as we went through each page of this chapter.
I often saw that after I have learnt a new concept in a chapter, it would not become intuitive immediately. I would understand the concepts, understand the related theorems, understand each step of the proofs, solve exercise problems, know how to apply the theorems when needed, and yet I could not "feel" them. I wanted to not just understand the concepts but I also wanted to "feel" the concepts like the way I could feel algebra, calculus, computer programming, etc. In the initial days, I wondered if I was too old to develop good intuition for all these new and highly sophisticated concepts.
Despite always feeling that all these concepts were too technical and quite unintuitive, I kept going. I kept hosting these discussions with a frequency of about 3-5 days every week. We continued discussing the various chapters and the proofs in them. And then suddenly one day while reading chapter 4, something interesting happened. As we were employing Dirichlet products to obtain some useful results, I realised that the concept of Dirichlet products which once felt so foreign two chapters earlier, now felt completely intuitive. I could see different functions being equivalent to Dirichlet products intuitively and effortlessly. Dirichlet products felt no more alien than, say, arithmetic multiplication. I could "feel" it now. It was a great feeling. I realised that sometimes it might take a few additional chapters of reading and using those concepts over and over again before they really begin to feel intuitive.
In this section, I will pick three interesting concepts from different parts of the book to provide a glimpse of what the journey has been like. These three things occur in the book again and again and play a very important role in several chapters of the book. Of course, it goes without saying that there are many interesting concepts in the book and many of them may be more important than the ones I am about to show below.
For any positive integer \( n, \) the Möbius function \( \mu(n) \) is defined as follows: \[ \mu(1) = 1; \] If \( n > 1, \) write \( n = p_1^{a_1} \dots p_k^{a_k} \) (prime factorisation). Then \begin{align*} \mu(n) & = (-1)^k \text{ if } a_1 = a_2 = \dots = a_k = 1, \\ \mu(n) & = 0 \text{ otherwise}. \end{align*} If \( n \ge 1, \) we have \[ \sum_{d \mid n} \mu(d) = \begin{cases} 1 & \text{ if } n = 1, \\ 0 & \text{ if } n > 1. \end{cases} \]
I was unfamiliar with this function prior to reading the book. It felt like a nice little cute function initially but as we went through more chapters, it soon became clear that this function plays a major role in analytic number theory.
As a simple example, we will soon see in this post that the Euler's totient function can be expressed as a Dirichlet product of the Möbius function and the arithmetical function \( N(n) = n. \)
As a more sophisticated example, the Dirichlet series with coefficients as the Möbius function is the multiplicative inverse of the Riemann zeta function, i.e., if \( s = \sigma + it \) is a complex number with its real part \( \sigma > 1, \) we have \[ \sum_{n=1}^{\infty} \frac{\mu(n)}{n^s} = \frac{1}{\zeta(s)}. \] This immediately shows that \( \zeta(s) \ne 0 \) for \( \sigma > 1. \)
If \( f \) and \( g \) are two arithmetical functions, their Dirichlet product \( f * g \) is defined as: \[ (f * g)(n) = \sum_{d \mid n} f(d) g\left( \frac{n}{d} \right). \] Dirichlet products appear to pop up magically at various places in number theory. Here is a simple example: \[ \varphi(n) = \sum_{d \mid n} \mu(d) \frac{n}{d}. \] Therefore in the notation of Dirichlet products, the above equation can also be written as \[ \varphi = \mu * N \] where \( N \) represents the arithmetical function \( N(n) = n \) for all \( n. \)
For complex numbers \( s = \sigma + it, \) the Hurwitz zeta function \( \zeta(s, a) \) is initially defined for \( \sigma > 1 \) as \[ \zeta(s, a) = \sum_{n=0}^{\infty} \frac{1}{(n + a)^s} \] where \( a \) is a fixed real number, \( 0 < a < 1. \) Then by analytic continuation, it is defined for \( \sigma \le 1 \) as \[ \zeta(s, a) = \Gamma(1 - s)I(s, a) \] where \( \Gamma \) represents the gamma function \[ \Gamma(s) = \int_0^{\infty} x^{s - 1} e^{-x} \, dx \] defined for \( \sigma > 0 \) and also defined, by analytic continuation, for \( \sigma \le 0 \) except for \( \sigma = 0, -1, -2, \dots \) (the nonpositive integers) and \( I(s, a) \) is defined by the contour integral \[ I(s, a) = \frac{1}{2\pi i} \int_C \frac{z^{s-1} e^{az}}{1 - e^z} \, dz \] where \( 0 < a \le 1 \) and the contour \( C \) is a loop around the negative real axis composed of three parts \( C_1, \) \( C_2, \) and \( C_3 \) such that for \( c < 2\pi, \) we have \( z = re^{-\pi i} \) on \( C_1 \) and \( z = re^{\pi i} \) on \( C_3 \) as \( r \) varies from \( c \) to \( +\infty, \) and \( z = ce^{i \theta} \) on \( C_2, \) \( -\pi \le \theta \le \pi. \)
Now admittedly, the definition or the analytic continuation of Hurwitz zeta function may seem very heavy and obscure to the uninitiated and it is indeed quite heavy. It takes 6 pages in chapter 12 to build the prerequisite concepts before we arrive at this definition. It is evident that this definition uses other concepts like the gamma function, a specific contour integral, etc. and it is only natural to expect that one has to gain sufficient expertise with the gamma function and contour integrals before the Hurwitz zeta function begins to feel intuitive.
But once we have established the analytic continuation of the Hurwitz zeta function, many insightful facts about the Riemann zeta function follow readily. It is easy to see that the Riemann zeta function can be defined in terms of the Hurwitz zeta function as \[ \zeta(s) = \zeta(s, 1) = \sum_{n=1}^{\infty} \frac{1}{n^s}. \] Yes, the \( \zeta \) symbol is overloaded: \( \zeta(s, a) \) is the Hurwitz zeta function whereas \( \zeta(s) \) is the Riemann zeta function. This relationship between the Riemann zeta function and the Hurwitz zeta function along with the analytic continuation of the Hurwitz zeta function opens new doors into the wonderful world of complex numbers and let us obtain beautiful and profound facts about the Riemann zeta function such as the fact that it has zeros at negative even integers, i.e., \( \zeta(n) = 0 \) for \( n = -2, -4, -6, \dots \) and the fact that \( \zeta(0) = -\frac{1}{2} \) and \( \zeta(-1) = -\frac{1}{12} \) and so on.
I believe beautiful results like these obtained by digging deep into complex analysis are what makes the study of analytic number theory so rewarding.
The next meeting is coming up today in a few hours. Are we planning anything special for the 100th meeting?
I think the 100th meeting is a significant milestone in our journey of understanding the beautiful and interesting gems hidden away in the subject of analytic number theory. This milestone has been possible only due to the sustained curiousity and eagerness among the members of the group to learn a significant area of mathematics and learn it well. We have reached this milestone successfully due to the passion and love for mathematics that drive the regular members to join these meetings and go through a few pages of the book everyday. In these meetings, we have read 12 chapters consisting of over 250 pages so far. Many of us knew nothing about analytic number theory merely five months ago and now we can appreciate the Riemann zeta function at a deeper level. We now understand what the Riemann hypothesis really means. This has been a great journey so far.
Despite being a significant milestone and cause for celebration, we are going to keep our 100th meeting fairly simple. We will continue where we left off yesterday. Today we have some more relationships between the gamma function and the Riemann zeta function to go through, so that is what we will do. We will also show that \( \zeta(0) = -\frac{1}{2} \) and \( \zeta(-1) = -\frac{1}{12} \) using the analytic continuation of the Hurwitz zeta function today.
If this blog post was fun for you and you would like to join our meetups, please go through this page to get the meeting link and join us.
]]>We summarise with this, the most remarkable formula in mathematics: \[ e^{i \theta} = \cos \theta + i \sin \theta. \] This is our jewel.
We may relate the geometry to the algebra by representing complex numbers in a plane; the horizontal position of a point is \( x, \) the vertical position of a point is \( y. \) We represent every complex number, \( x + iy. \) Then if the radial distance to this point is called \( r \) and the angle is called \( \theta, \) the algebraic law is that \( x + iy \) is written in the form \( r, e^{i \theta} \) where the geometrical relationships between \( x \) \( y, \) \( r, \) and \( \theta \) are as shown. This, then, is the unification of algebra and geometry.
See the bottom of the page at https://www.feynmanlectures.caltech.edu/I_22.html for the above excerpt.
]]>Important numbers in the proof: \[ 0, \quad \underbrace{[y]}_{=\,m}, \quad y, \quad \underbrace{[y] + 1}_{=\,m + 1}, \quad \underbrace{[x]}_{=\,k}, \quad x. \] Splitting the definite integral: \[ \int_y^x f(t)\,dt = \int_{y}^{[y] + 1} f(t)\,dt + \underbrace{\int_{[y] + 1}^{[y] + 2} f(t)\,dt + \dots + \int_{[x] - 1}^{[x]} f(t)\,dt}_{=\,\int_{[y] + 1}^{[x]} f(t)\, dt} + \int_{[x]}^{x} f(t)\,dt. \] Using the more convenient variables \( m \) and \( k, \) we get: \[ \int_y^x f(t)\,dt = \int_m^{m + 1} f(t)\,dt + \underbrace{\int_{m + 1}^{m + 2} f(t)\,dt + \dots + \int_{k - 1}^{k} f(t)\,dt}_{=\,\int_{m + 1}^{k} f(t)\, dt} + \int_{k}^{x} f(t)\,dt. \]
\begin{align*} \int_{m + 1}^{k} [t] f'(t) dt & = \int_{m + 1}^{m + 2} [t] f'(t) dt + \int_{m + 2}^{m + 3} [t] f'(t) dt + \dots + \int_{k - 1}^{k} [t] f'(t) dt \\ & = \begin{aligned}[t] & (m + 2) f(m + 2) - (m + 1) f(m + 1) - f(m + 2) \\ + & (m + 3) f(m + 3) - (m + 2) f(m + 2) - f(m + 3) \\ & \dots \\ + & (k) f(k) - (k - 1) f(k - 1) - f(k) \end{aligned} \\ & = kf(k) - (m + 1)f(m + 1) - \sum_{n=m + 2}^{k} f(n) \\ & = kf(k) - mf(m + 1) - f(m + 1) - \sum_{n=m + 2}^{k} f(n) \\ & = kf(k) - mf(m + 1) - \sum_{n=m + 1}^{k} f(n) \\ & = kf(k) - mf(m + 1) - \sum_{y < n \le x} f(n). \end{align*}
\begin{align*} \sum_{y < n \le x} f(n) & = - \int_{m + 1}^k [t] f'(t) \, dt + k f(k) - m f(m + 1) \\ & = \begin{aligned}[t] & \left( - \int_y^{m + 1} [t] f'(t) \, dt - \int_{m + 1}^k [t] f'(t) \, dt - \int_k^x [t] f'(t) \, dt \right) \\ & + f(k) - m f(m + 1) + \int_y^{m + 1} [t] f'(t) \, dt + \int_k^x [t] f'(t) \, dt \end{aligned} \\ & = - \int_y^x [t] f'(t) \, dt + k f(k) - m f(m + 1) + \int_y^{m + 1} m f'(t) \, dt + \int_k^x k f'(t) \, dt \\ & = - \int_y^x [t] f'(t) \, dt + k f(k) - m f(m + 1) + \biggl( m f(m + 1) - m f(y) \biggr) + \biggl( k f(x) - k f(k) \biggr) \\ & = - \int_y^x [t] f'(t) \, dt + k f(x) - m f(y). \end{align*}
Integration by parts: \[ \int uv \, dt = u \int v \, dt - \int u' \left( \int v \, dt \right) \, dt. \] \[ \int_y^x t f'(t) \, dt = \left. \left( t f(t) - \int f(t) \, dt \right) \right|_y^x = x f(x) - y f(y) - \int_y^x f(t) \, dt. \] Final step of the proof: \begin{align*} \sum_{y < n \le x} f(n) & = -\int_y^x [t] f'(t) \, dt + k f(x) - m f(y) \\ & = \begin{aligned}[t] & -\int_y^x [t] f'(t) \, dt + [x] f(x) - [y] f(y) \\ & + \underbrace{ \left( \int_y^x t f'(t) \, dt - x f(x) + y f(y) + \int_y^x f(t) \, dt \right)}_{0 \text{ by above definite integral}} \end{aligned} \\ & = \int_y^x f(t) \, dt + \int_y^x (t - [t]) f'(t) \, dt + f(x)([x] - x) - f(y)([y] - y). \end{align*}
Splitting definite integral: \begin{align*} & \int_1^{\infty} f(t) \, dt = \int_1^{x} f(t) \, dt + \int_x^{\infty} f(t) \, dt \\ & \iff \int_1^{\infty} f(t) \, dt - \int_x^{\infty} f(t) \, dt = \int_1^x f(t) \, dt. \end{align*} Solving improper integral: \[ \int_x^{\infty} \frac{1}{t^2} \, dt = \lim_{b \to \infty} \int_x^b \frac{1}{t^2} dt = \lim_{b \to \infty} \frac{-1}{t} \Biggr|_x^b = \left( \lim_{b \to \infty} \frac{-1}{b} \right) + \frac{1}{x} = 0 + \frac{1}{x} = \frac{1}{x}. \]
Definition of Euler's constant: \[ C = \lim_{n \to \infty} \left( 1 + \frac{1}{2} + \frac{1}{3} + \dots + \frac{1}{n} - \log n \right) = \lim_{x \to \infty} \left( \sum_{n \le x} \frac{1}{n} - \log x \right). \] We begin with \[ \sum_{n \le x} \frac{1}{n} = \log x + \underbrace{1 - \int_1^{\infty} \frac{t - [t]}{t^2} \, dt}_{\text{We will show below that this is \( C \)}} + O\left( \frac{1}{x} \right). \] Rearranging the terms, we get \[ \sum_{n \le x} \frac{1}{n} - \log x = 1 - \int_1^{\infty} \frac{t - [t]}{t^2} \, dt + O\left( \frac{1}{x} \right). \] Using the definition of \( C, \) we get \begin{align*} C & = \lim_{x \to \infty} \left( \sum_{n \le x} \frac{1}{n} - \log x \right) \\ & = \lim_{x \to \infty} \left( 1 - \int_1^{\infty} \frac{t - [t]}{t^2} \, dt + O\left( \frac{1}{x} \right) \right) \\ & = 1 - \int_1^{\infty} \frac{t - [t]}{t^2} \, dt. \end{align*}
\[ \int_1^x \frac{dt}{t^s} = \frac{t^{-s + 1}}{-s + 1} \Biggr|_1^x = \frac{t^{1 - s}}{1 - s} \Biggr|_1^x = \frac{x^{1 - s}}{1 - s} - \frac{1}{1 - s}. \] \[ \int_1^x \frac{t - [t]}{t^{s + 1}} \, dt = \int_1^{\infty} \frac{t - [t]}{t^{s + 1}} \, dt - \int_x^{\infty} \frac{t - [t]}{t^{s + 1}} \, dt = \int_1^{\infty} \frac{t - [t]}{t^{s + 1}} \, dt + \underbrace{\frac{1}{s} O\left( x^{-s}\right)}_{\text{explained below}}. \] \[ 0 \le \int_x^{\infty} \frac{t - [t]}{t^{s + 1}} \, dt \le \int_x^{\infty} \frac{1}{t^{s + 1}} \, dt = \frac{-1}{st^s} \Biggr|_x^\infty = \frac{1}{sx^s} = \frac{1}{s} x^{-s}. \] \begin{align*} \sum_{n \le x} \frac{1}{n^s} & = \int_1^x \frac{dt}{t^s} - s \int_1^x \frac{t - [t]}{t^{s + 1}} + 1 - \frac{x - [x]}{x^s} \, dt \\ & = \frac{x^{1 - s}}{1 - s} - \frac{1}{1 - s} - s \int_1^{\infty} \frac{t - [t]}{t^{s + 1}} \, dt + 1 + O(x^{-s}). \end{align*}
Read on website | #mathematics | #number-theory | #book | #meetup
]]>\[ f(mn) = f(m) f(n) \text{ for all } m, n. \]
\[ I(n) = \begin{cases} 1 & \text{ if } n = 1, \\ 0 & \text{ if } n > 1. \end{cases} \]
\begin{align*} f(n)I(n) & = \begin{cases} 1 \cdot 1 & \text{ if } n = 1, \\ f(n) \cdot 0 & \text{ if } n > 1. \end{cases} \\ & = \begin{cases} 1 & \text{ if } n = 1, \\ 0 & \text{ if } n > 1. \end{cases} \\ & = I(n). \end{align*}
\[ \mu(1) = 1, \qquad \mu(p) = -1, \qquad \mu(p^2) = \mu(p^3) = \dots = 0. \]
\begin{align*} \sum_{d \mid p^a} \mu(d) f(d) f\left(\frac{p^a}{d}\right) & = \sum_{d = 1, p, p^2, \dots, p^a} \mu(d) f(d) f\left(\frac{p^a}{d}\right) \\ & = \begin{aligned}[t] & \mu(1) f(1) f\left( \frac{p^a}{1} \right) + \mu(p) f(p) f\left( \frac{p^a}{p} \right) \\ & + \underbrace{\mu(p^2) f(p^2) f\left( \frac{p^a}{p^2} \right) + \dots + \mu(p^a) f(p^a) f\left( \frac{p^a}{p^a} \right)}_{=\,0} \end{aligned} \\ & = \mu(1) f(1) f(p^a) + \mu(p) f(p) f(p^{a - 1}) \\ & = f(p^a) - f(p) f(p^{a - 1}). \end{align*}
\begin{align*} f(p^a) & = f(p)f(p^{a - 1}) \\ & = f(p)f(p)f(p^{a - 2}) \\ & = \dots \\ & = \underbrace{f(p)f(p)f(p) \dots f(p)}_{a \text{ times}} \\ & = \left( f(p) \right)^a. \end{align*}
\[ f(mn) = f(m)f(n) \text{ whenever } (m, n) = 1. \]
\[ f(p_1^{\alpha_1} p_2^{\alpha_2} \dots p_k^{\alpha_k}) = f(p_1^{\alpha_1}) f(p_2^{\alpha_2}) \dots f(p_k^{\alpha_k}). \]
\[ \varphi(n) = \sum_{d \mid n} \mu(d) \frac{n}{d} = \sum_{d \mid n} \mu(d) N\left(\frac{n}{d}\right) = (\mu * N)(n). \]
Let \( f \) be multiplicative. We want to show that \[ \sum_{d \mid n} \mu(d) f(d) = \prod_{p \mid n} (1 - f(p)). \] Note the following: \[ g(n) = \sum_{d \mid n} \mu(d) f(d) = \sum_{d \mid n} (\mu f) (d) u\left( \frac{n}{d} \right) = (\mu f) * u. \] The functions \( \mu \) and \( f \) are multiplicative. Thus \( \mu f \) is multiplicative. Thus \( (\mu f) * u \) is multiplicative. Therefore \[ g(n) = g(p_1^{a_1} p_2^{a_2} \dots p_k^{a_k}) = g(p_1^{a_1}) g(p_2^{a_2}) \dots g(p_k^{a_k}). \] But \begin{align*} g(p_i^{a_i}) & = \sum_{d \mid p_i^{a_i}} \mu(d) f(d) \\ & = \mu(1) f(1) + \mu(p_i) f(p_i) + \underbrace{\mu(p_i^2) f(p_i^2) + \dots + \mu(p_i^{a_i}) f(p_i^{a_i})}_{=\,0} \\ & = 1 - f(p). \end{align*} From the two equations above, we get \begin{align*} g(n) & = g(p_1^{a_1}) g(p_2^{a_2}) \dots g(p_k^{a_k}) \\ & = (1 - f(p_1)) (1 - f(p_2)) \dots (1 - f(p_k)) \\ & = \prod_{p \mid n} (1 - f(p)). \end{align*}
\begin{align*} A(x)B(x) & = \left( \sum_{n=0}^{\infty} a(n) x^n \right) \left( \sum_{n=0}^{\infty} b(n) x^n \right) \\ & = \left( a(0) + a(1)x + a(2)x^2 + \dots \right) \left( b(0) + b(1)x + b(2)x^2 + \dots \right) \\ & = a(0)b(0) + \Bigl( a(0)b(1) + a(1)b(0) \Bigr) x + \Bigl( a(0)b(2) + a(1)b(1) + a(2)b(0) \Bigr) x^2 + \dots \\ & = \sum_{k=0}^0 a(k)b(n - k) + \sum_{k=0}^1 a(k)b(1 - k)x + \sum_{k=0}^2 a(k)b(2 - k)x^2 + \dots \\ & = \sum_{n=0}^{\infty} \sum_{k=0}^n a(k)b(n - k). \end{align*}
\[ A(x)B(x) = \sum_{n=0}^{\infty} \underbrace{\left\{ \sum_{k=0}^{n} a(k) b(n - k) \right\}}_{c(n)} x^n. \] \[ B(x)A(x) = \sum_{n=0}^{\infty} \underbrace{\left\{ \sum_{k=0}^{n} a(n - k) b(k) \right\}}_{c'(n)} x^n. \] \[ c(3) = a(0)b(3) + a(1)b(2) + a(2)b(1) + a(3)b(0). \] \[ c'(3) = a(3)b(0) + a(2)b(1) + a(1)b(2) + a(0)b(3). \]
\[ A(x)\Bigl(B(x) + C(x)\Bigr) = A(x)B(x) + A(x)C(x). \] \[ \Bigl(B(x) + C(x)\Bigr)A(x) = B(x)A(x) + C(x)A(x). \] \begin{align*} A(x)\Bigl(B(x) + C(x)\Bigr) & = \left( \sum_{n=0}^{\infty} a(n) x^n \right) \left( \sum_{n=0}^{\infty} \Bigl( b(n) + c(n) \Bigr) x^n \right) \\ & = \sum_{n=0}^{\infty} \Bigl\{ \sum_{k=0}^{n} a(k) \Bigl( b(n - k) + c(n - k) \Bigr) \Bigr\} x^n. \end{align*} \[ A(x)B(x) + A(x)C(x) = \sum_{n=0}^{\infty} \sum_{k=0}^n a(k) b(n - k) x^n + \sum_{n=0}^{\infty} \sum_{k=0}^n a(k) c(n - k) x^n. \]
\begin{align*} A(x)B(x) & = \sum_{n=0}^{\infty} \Bigl( \sum_{k=0}^{n} a(k) b(n - k) \Bigr) x^n \\ & = \Bigl( a(0) b(0) \Bigr) x^0 + \Bigl( a(0) b(1) + a(1) b(0) \Bigr) x^1 + \Bigl( a(0) b(2) + a(1) b(1) + a(2) b(0) \Bigr) x^2 + \dots \\ & = 1. \end{align*}
\begin{align*} A(x) & = 1 + ax + (ax)^2 + (ax)^3 + \dots, \\ B(x) & = 1 - ax. \end{align*} \begin{align*} A(x) B(x) & = \Bigl( 1 + ax + (ax)^2 + (ax)^3 + \dots \Bigr) (1 - ax) \\ & = \Bigl( 1 + ax + (ax)^2 + (ax)^3 + \dots \Bigr) - \Bigl( (ax) - (ax)^2 - (ax)^3 - \dots \Bigr) = 1. \end{align*}
\[ f_p(x) = \sum_{n=0}^{\infty} f(p^n) x^n = f(1) + f(p) x + f(p^2) x^2 + f(p^3) x^3 + \dots \]
\begin{align*} f(n) & = f(p_1^{a_1} p_2^{a_2} \dots p_k^{a_k}) = f(p_1^{a_1}) f(p_2^{a_2}) \dots f(p_k^{a_k}), \\ \\ g(n) & = g(p_1^{a_1} p_2^{a_2} \dots p_k^{a_k}) = g(p_1^{a_1}) g(p_2^{a_2}) \dots g(p_k^{a_k}). \\ \end{align*}
\begin{align*} \mu_p(x) = \sum_{n=0}^{\infty} \mu(p^n) x^n & = \mu(1) + \mu(p) x + \mu(p^2) x^2 + \mu(p^3) x^3 + \dots \\ & = 1 - x + 0 + 0 + \dots \\ & = 1 - x. \end{align*}
\[ A(x) = \sum_{n=0}^{\infty} a(n) x^n, \quad B(x) = \sum_{n=0}^{\infty} b(n) x^n, \quad A(x) B(x) = \sum_{n=0}^{\infty} \underbrace{\sum_{k=0}^n a(k) b(n - k)}_{c(n)} x^n. \]
\[ (f * g)_p(x) = f_p(x) g_p(x). \] \[ f_p(x) = \sum_{n=0}^{\infty} f(p^n) x^n, \quad g_p(x) = \sum_{n=0}^{\infty} g(p^n) x^n, \quad f_p(x) g_p(x) = \sum_{n=0}^{\infty} \sum_{k=0}^n f(p^k) g(p^{n-k}) x^n. \] \[ h = f * g = \sum_{d \mid n} f(d) g\left( \frac{n}{d} \right). \] \[ h_p(x) = \sum_{n=0}^{\infty} h(p^n) x^n = \sum_{n=0}^{\infty} \sum_{d \mid p^n} f(d) g\left( \frac{p^n}{d} \right) x^n = \sum_{n=0}^{\infty} \sum_{k=0}^{n} f(p^k) g(p^{n-k}) x^n. \]
Some steps of Example 1: \[ I(n)= \mu^2(n) * \lambda(n) \implies I_p(x) = \mu_p^2(x) \lambda_p(x) \] \[ I_p(x) = \mu_p^2(x) \lambda_p(x) \iff 1 = \mu_p^2(x) \cdot \frac{1}{1 + x} \iff \mu_p^2(x) = 1 + x. \]
Some steps of Example 2: \begin{align*} \frac{1}{1 - p^{\alpha}} \cdot \frac{1}{1 - x} & = \frac{1}{1 - x - p^{\alpha}x + p^{\alpha}x^2} \\ & = \frac{1}{1 - (1 + p^{\alpha})x + p^{\alpha}x^2} \\ & = \frac{1}{1 - \sigma_{\alpha}(p)x + p^{\alpha}x^2}. \end{align*} Note that \( \sigma_{\alpha}(n) = \sum_{d\,\mid\,n} d^{\alpha}, \) so \[ \sigma_{\alpha}(p) = \sum_{d\,\mid\,p} d^{\alpha} = 1^{\alpha} + p^{\alpha} = 1 + p^{\alpha}. \]
Some steps of Example 3: Showing That \( f(n) = 2^{\nu(n)} \) is multiplicative: \[ f(n) = 2^{\nu(n)}. \] \[ f(p_1^{\alpha_1} p_2^{\alpha_2} \dots p_k^{\alpha_k}) = 2^{\nu(p_1^{\alpha_1} p_2^{\alpha_2} \dots p_k^{\alpha_k})} = 2^k. \] \[ f(p_1^{\alpha_1}) f(p_2^{\alpha_2}) \dots f(p_k^{\alpha_k}) = 2^{\nu(p_1^{\alpha_1})} 2^{\nu(p_2^{\alpha_2})} \dots 2^{\nu(p_k^{\alpha_k})} = \underbrace{2 \cdot 2 \cdot \dots \cdot 2}_{k \text{ times}}. = 2^k. \]
\[ (f + g)' = f' + g'. \] \[ (fg)' = f'g + fg'. \] \[ \left( f^{-1} \right)' = \frac{-f'}{f^2} = -f' \cdot (f \cdot f)^{-1}. \]
\[ f'(n) = f(n) \log n. \] \[ (f + g)' = f' + g'. \] \[ (f * g)' = f' * g + f * g'. \] \[ \left( f^{-1} \right)' = -f' * (f * f)^{-1}. \]
Read on website | #mathematics | #number-theory | #book | #meetup
]]>This page contains an archive of notes from the book Introduction to Analytic Number Theory by Tom M. Apostol (1976).
Note that this set of notes is not meant to be a systematic exposition of analytic number theory. Instead this is just a collection of examples that illustrate some of the theorems in the reference textbook and intermediate steps that are not explicitly expressed in the book. These boards were used to aid the discussions during book discussion meetings. As a result, the content of these boards is informal in nature and is not intended to be a substitute for the book or the actual discussion meetings.
If you find any mistakes in the content of the board files, please create a new issue or send a pull request.
More notes coming soon! We have all the meeting notes safely archived. Just need to format them and publish them here.
Read on website | #mathematics | #number-theory | #book | #meetup
]]>The following content on this page is an archive of the content as it appeared on the last day of meeting for this book.
Meeting time: 17:00 UTC from Tuesday to Friday, usually.^{†}
Meeting duration: 40 minutes.
Meeting link: bit.ly/spzoom2
Meeting log: 120 meetings
Reference Book: Introduction to Analytic Number Theory by Tom M. Apostol (1976)
Chapter notes: Notes
Started: 05 Mar 2021
Ended: 01 Oct 2021
† There are some exceptions to this schedule occasionally. Join our channel to receive schedule updates.
The primary reference book for these meetings is Introduction to Analytic Number Theory written by Tom M. Apostol. Admittedly, the book is quite expensive but you may find a relatively cheap paperback (softcover) copy on some websites.
These meetings are hosted by Susam and attended by some members of
#math
and #algorithms
channels of Libera
IRC network as well as by some members
from Hacker News.
You are welcome to join these meetings anytime. If you are concerned that the meetings may not make sense if you join when we are in the middle of a chapter, please free to talk to us about it in the group channel. I can recommend the next best time to begin joining the meetings. Usually, it would be when we begin reading a new section or chapter that is fairly self-contained and does not depend a lot on material we have read previously.
Read on website | #mathematics | #number-theory | #book | #meetup
]]>One striking characteristic of Grothendieck's mode of thinking is that it seemed to rely so little on examples. This can be seen in the legend of the so-called "Grothendieck prime". In a mathematical conversation, someone suggested to Grothendieck that they should consider a particular prime number. "You mean an actual number?" Grothendieck asked. The other person replied, yes, an actual prime number. Grothendieck suggested, “All right, take 57.”]]>
(year % 4 == 0 && year % 100 != 0) || year % 400 == 0
It came as a surprise to me. Prior to reading this, I did not know that centurial years are not leap years except for those centurial years that are also divisible by 400. Until then, I always incorrectly thought that all years divisible by 4 are leap years. I have witnessed only one centurial year, namely the year 2000, which happens to be divisible by 400. As a result, the year 2000 proved to be a leap year and my misconception remained unchallenged for another few years until I finally came across the above test in K&R.
Now that I understand that centurial years are not leap years unless
divisible by 400, it is easy to confirm this with the
Unix cal
command. Enter cal 1800
or cal 1900
and we see calendars of non-leap years.
But enter cal 2000
and we see the calendar of a leap
year.
By the way, the following leap year test is equally effective:
year % 4 == 0 && (year % 100 != 0 || year % 400 == 0)
Update: In the
comments section,
Thaumasiotes explains why both tests work. Let me take the liberty
of elaborating that comment further with a truth table. We use the
notation A
, B
, and C
,
respectively, for the three comparisons in the above expressions.
Then the two tests above can be expressed as the following boolean expressions:
(A && B) || C
A && (B || C)
Now normally these two boolean expressions are not equivalent. The truth table below shows this:
A |
B |
C |
(A && B) || C |
A && (B || C) |
---|---|---|---|---|
F | F | F | F | F |
F | F | T | T | F |
F | T | F | F | F |
F | T | T | T | F |
T | F | F | F | F |
T | F | T | T | T |
T | T | F | T | T |
T | T | T | T | T |
We see that there are two cases where the last two columns differ.
Therefore indeed the two boolean expressions are not equivalent.
The two cases where the boolean expressions yield different results
occur when A
is false and C
is true. But
these cases are impossible! If A
is false
and C
is true, it means we have year % 4 !=
0
and year % 400 == 0
which is impossible.
If year % 400 == 0
is true, then year % 4 ==
0
must also hold true. In other words, if C
is
true, A
must also be true. Therefore, the two cases
where the last two columns differ cannot occur and may be ignored.
The last two columns are equal in all other cases and that is why
the two tests we have are equivalent.
Read on website | #c | #programming | #technology | #book | #mathematics
]]>A fascinating result that appears in linear algebra is the fact that the set of real numbers \( \mathbb{R} \) is a vector space over the set of rational numbers \( \mathbb{Q}. \) This may appear surprising at first but it is easy to show that it is indeed so by checking that all eight axioms of vector spaces hold good:
Commutativity of vector addition:
\( x + y = y + x \) for all \( x, y \in \mathbb{R}. \)
Associativity of vector addition:
\( x + (y + z) = (x + y) + z \) for all \( x, y, z \in
\mathbb{R}. \)
Existence of additive identity vector:
We have \( 0 \in \mathbb{R} \) such that \( x + 0 = x \) for all
\( x \in \mathbb{R}. \)
Existence of additive inverse vectors:
There exists \( -x \in \mathbb{R} \) for every \( x \in
\mathbb{R} such that \( x + (-x) = 0. \)
Associativity of scalar multiplication:
\( a(bx) = (ab)x \) for all \( a, b \in \mathbb{Q} \) and all \(
x \in \mathbb{R}. \)
Distributivity of scalar multiplication over vector
addition:
\( a(x + y) = ax + by \) for all \( a \in \mathbb{Q} \) and all
\( x, y \in \mathbb{R}. \)
Distributivity of scalar multiplication over scalar
addition:
\( (a + b)x = ax + bx \) for all \( a, b \in \mathbb{Q} \) and
all \( x \in \mathbb{R}. \)
Existence of scalar multiplicative identity:
We have \( 1 \in \mathbb{Q} \) such that \( 1 \cdot x = x \) for
all \( x \in \mathbb{R}. \)
This shows that the set of real numbers \( \mathbb{R} \) forms a vector space over the field of rational numbers \( \mathbb{Q}. \) Another quick way to arrive at this fact is to observe that \( \mathbb{Q} \subseteq \mathbb{R}, \) that is, \( \mathbb{Q} \) is a subfield of \( \mathbb{R}. \) Any field is a vector space over any of its subfields, so \( \mathbb{R} \) must be a vector space over \( \mathbb{Q}. \)
We can also show that \( \mathbb{R} \) is an infinite dimensional vector space over \( \mathbb{Q}. \) Let us assume the opposite, i.e., \( \mathbb{R} \) is finite dimensional. Let \( r_1, \dots, r_n \) be the basis for this vector space. Therefore for each \( r \in \mathbb{R}, \) we have unique \( q_1, \dots, q_n \in \mathbb{Q} \) such that \( r = q_1 r_1 + \dots + q_n r_n. \) Thus there is a bijection between \( \mathbb{Q}^n \) and \( \mathbb{R}. \) This is a contradiction because \( \mathbb{Q}^n \) is countable whereas \( \mathbb{R} \) is uncountable. Therefore \( \mathbb{R} \) must be an infinite dimensional vector space over \( \mathbb{Q}. \)
Here is an interesting problem related to vector spaces that I came across recently:
Define two periodic functions \( f \) and \( g \) from \( \mathbb{R} \) to \( \mathbb{R} \) such that their sum \( f + g \) is the identity function. The axiom of choice is allowed.
A function \( f \) is periodic if there exists \( p \gt 0 \) such that \( f(x + p) = f(x) \) for all \( x \) in the domain.
If you want to think about this problem, this is a good time to pause and think about it. There are spoilers ahead.
The axiom of choice is equivalent to the statement that every vector space has a basis. Since the set of real numbers \( \mathbb{R} \) is a vector space over the set of rational numbers \( \mathbb{Q}, \) there must be a basis \( \mathcal{H} \subseteq \mathbb{R} \) such that every real number \( x \) can be written uniquely as a finite linear combination of elements of \( \mathcal{H} \) with rational coefficients, that is, \[ x = \sum_{a \in \mathcal{H}} x_a a \] where each \( x_a \in \mathbb{Q} \) and \( \{ a \in \mathcal{H} \mid x_a \ne 0 \} \) is finite. The set \( \mathcal{H} \) is also known as the Hamel basis.
In the above expansion of \( x, \) we use the notation \( x_a \) to denote the rational number that appears as the coefficient of the basis vector \( a. \) Therefore \( (x + y)_{a} = x_a + y_a \) for all \( x, y \in \mathbb{R} \) and all \( a \in \mathcal{H}. \)
We know that \( b_a = 0 \) for distinct \( a, b \in \mathcal{H} \) because \( a \) and \( b \) are basis vectors. Thus \( (x + b)_{a} = x_a + b_a = x_a + 0 = x_a \) for all \( x \in \mathbb{R} \) and distinct \( a, b \in \mathcal{H}. \) This shows that a function \( f(x) = x_a \) is a periodic function with period \( b \) for any \( a \in \mathcal{H} \) and any \( b \in \mathcal{H} \setminus \{ a \}. \)
Let us define two functions: \begin{align*} g(x) & = \sum_{a \in \mathcal{H} \setminus \{ b \}} x_a a, & h(x) & = x_b b. \end{align*} where \( b \in \mathcal{H} \) and \( x \in \mathbb{R}. \) Now \( g(x) \) is a periodic function with period \( b \) for any \( b \in \mathcal{H} \) and \( h(x) \) is a periodic function with period \( c \) for any \( c \in \mathcal{H} \setminus \{ b \}. \) Further, \[ g(x) + h(x) = \left( \sum_{a \in \mathcal{H} \setminus \{ b \}} x_a a \right) + x_b b = \sum_{a \in \mathcal{H}} x_a a = x. \] Thus \( g(x) \) and \( h(x) \) are two periodic functions such that their sum is the identity function.
Everytime I travel to the US, one thing that troubles me a little is having to convert temperature from the Celsius scale to the Fahrenheit scale and vice versa. The exact conversion formulas are: \begin{align*} f & = \frac{9c}{5} + 32, \\ c & = \frac{5(f - 32)}{9} \end{align*} where \( f \) represents the temperature value in Fahrenheit and \( c \) represents the temperature value in Celsius.
While the formulas above are accurate, they are not very convenient to mentally figure what I need to set a room thermostat in Fahrenheit scale to if I want to keep the room at, say, 25 °C. For this particular case, I have memorised that 25 °C is 77 °F. This combined with the fact that every 5 °C interval corresponds to an interval of 9 °F, it is easy to mentally compute that 20 °C is 68 °F or 22.5 °C is 72.5 °F. It would still be nice to find an easy way to mentally convert any arbitrary temperature in one scale to the other scale.
In my last trip to the US, I decided to devise a few approximation methods to convert temperature from the Fahrenheit scale to the Celsius scale and vice versa. I arrived at two methods: one to convert temperature value in Fahrenheit to Celsius and another to convert from Fahrenheit to Celsius. Both these methods are based on the exact conversion formulas but they sacrifice accuracy a little bit in favour of simplifying the computations, so that they can be performed mentally.
Before we dive into the refined approximation methods I have arrived at, let us first see a very popular method that obtains a crude approximation of the result of temperature conversion from °C to °F and vice versa pretty quickly.
To go from °C to °F, we perform the following two steps:
To go from °F to °C, we perform the inverse:
We arrive at the above methods by approximating 9/5 and 32 in the exact conversion formulas with 2 and 30, respectively. These methods can be performed mentally quite fast but this speed of mental calculation comes at the cost of accuracy. That's why I call them crude approximation methods.
The first method converts 10 °C exactly to 50 °F without any error. But then it introduces an error of 1 °F for every 5 °C interval. For example, the error is 3 °F for 25 °C and 18 °F for 100 °C.
Similarly, the second method converts 50 °F exactly to 10 °C without any error. But it introduces an error of 0.5 °C for every 9 °F interval. For example, the error is 1.5 °C for 77 °C and 9 °C for 212 °F.
Let us do a few examples to see how well the crude approximation methods work. Let us say, we want to convert 24 °C to °F.
The exact value for 24 °C is 75.2 °F. This approximation method overestimated the actual temperature in Fahrenheit by 2.8 °F.
Let us now convert 75 °F to °C.
The exact value for 75 °F is 23.89 °C. This approximation method underestimated the actual temperature in Celsius by 1.39 °C.
Can we do better?
This section presents the refined approximation methods that I have arrived at. They are a little slower to perform mentally than the crude approximation methods but they are more accurate.
To keep the methods convenient enough to perform mentally, we work with integers only. We always start with an integer value in Celsius or Fahrenheit. The result of conversion is also an integer. If a fraction arises in an intermediate step, we discard the fractional part. For example, if a step requires us to calculate one-tenth of a number, say, 25, we consider the result to be 2. Similarly, if a step requires us to halve the number 25, we consider the result to be 12. This is also known as truncated division or integer division or floor division.
To go from °C to °F, here is my quick three-step approximation method:
The approximation error due to this method does not exceed 1 °F in magnitude. In terms of Celsius, the approximation error does not exceed 0.56 °C. I believe this is pretty good if we are talking about setting the thermostat temperature.
To go from °F to °C, we perform a rough inverse of the above steps:
In fact, for integer temperature values between 32 °F (0 °C) and 86 °F (30 °C), the approximation error due to this method does not exceed 1.12 °C. Further, for integer temperature values between −148 °F (−100 °C) and 212 °F (100 °C), the approximation error does not exceed 1.89 °C. This is pretty good if we are talking about the weather.
Let us do a few examples to see how well the three-step methods above work. Let us say, we want to convert 24 °C to °F.
The exact value for 22 °C is 75.2 °F. The approximation method has underestimated the actual temperature in Fahrenheit by only 0.2 °F.
Let us now try to convert 75 °F to °C.
The exact value for 75 °F is 23.89 °C, so this approximation method overestimated the actual temperature in Celsius by 0.11 °C only.
If you were looking only for quick methods to convert temperature values in Fahrenheit to Celsius and vice versa, this is all you need to know. You may skip the remaining post unless you want to know why these methods work.
In this section, we will see why the refined approximation methods work so well.
The method to convert temperature value from Celsius to Fahrenheit is equivalent to \[ \overset{\approx}{f} = 2 \left(c - \left\lfloor \frac{c}{10} \right\rfloor \right) + 31 \] where \( c \) is the temperature value in Celsius and \( \overset{\approx}{f} \) is the approximate temperature value in Fahrenheit.
Here is a brief justification for this:
Now let us see how we arrive at the above approximate conversion formula. It's not too different from the exact conversion formula. The exact formula to convert temperature from Celsius to Fahrenheit is \[ f = \frac{9c}{5} + 32. \] This can be rewritten as \[ f = 2 \left(c - \frac{c}{10} \right) + 32. \] We don't want to deal with fractions, so we decide to approximate \( \frac{c}{10} \) in the above formula with \( \left\lfloor \frac{c}{10} \right\rfloor \) and get \[ \overset{\sim}{f} = 2 \left(c - \left\lfloor \frac{c}{10} \right\rfloor \right) + 32. \] where \( \overset{\sim}{f} \) is an approximation of the value in Fahrenheit. The floor division has the effect of potentially overestimating the final result by a value that is less than \( 2. \) This is the approximation error.
If we define the approximation error as \( \overset{\sim}{f} - f, \) then the approximation error lies in the half-open interval \( [0, 2). \) To ensure that the magnitude of the error never exceeds \( 1, \) i.e., to make the approximation error lie in the half-open interval \( [-1, 1), \) we subtract \( 1 \) from the above formula and get \[ \overset{\approx}{f} = 2 \left(c - \left\lfloor \frac{c}{10} \right\rfloor \right) + 31. \] This is the formula that the three-step method to convert temperature from Celsius to Fahrenheit is equivalent to.
The inverse method to convert temperature value from Fahrenheit to Celsius amounts to this formula: \[ \overset{\approx}{c} = \left\lfloor \frac{f - 31}{2} \right\rfloor + \left\lfloor \frac{f - 31}{20} \right\rfloor \] where \( f \) is the temperature value in Fahrenheit and \( \overset{\approx}{c} \) is the approximate temperature value in Celsius.
Here is a brief justification for this:
This is roughly an inverse of all the steps for converting a temperature value from Celsius to Fahrenheit. Let us see if this is close to the exact conversion formula \[ c = \frac{5(f - 32)}{9}. \]
It turns out that it is in fact close to the exact conversion formula as follows: \begin{align*} c & = \frac{5(f - 32)}{9} \\ & = 0.5556 (f - 32) \\ & \approx 0.55 (f - 31) \\ & = \frac{11 (f - 31)}{20} \\ & = \frac{f - 31}{2} + \frac{f - 31}{20} \\ & \approx \left\lfloor \frac{f - 31}{2} \right\rfloor + \left\lfloor \frac{f - 31}{20} \right\rfloor = \overset{\approx}{c} \end{align*}
Like we discussed earlier, the magnitude of the approximation error does not exceed 1.89 °C for integer values between −148 °F and 212 °F. The error here is a little bit more than the previous approximation method to convert temperature in Celsius to Fahrenheit but it is still small enough to give us a reasonably good estimate of what a temperature value in Fahrenheit would look like in Celsius when we are talking about the weather.
]]>Let us talk a little bit about integer underflow and undefined behaviour in C before we discuss the puzzle I want to share in this post.
#include <stdio.h>
int main()
{
int i;
for (i = 0; i < 6; i--) {
printf(".");
}
return 0;
}
This code invokes undefined behaviour. The value in variable
i
decrements to INT_MIN
after
|INT_MIN|
iterations. In the next iteration, there is a
negative overflow which is undefined for signed integers in C. On
many implementations though, INT_MIN - 1
wraps around
to INT_MAX
. Since INT_MAX
is not less than
6
, the loop terminates. With such implementations, this
code prints print |INT_MIN| + 1
dots. With 32-bit integers,
that amounts to 2147483649 dots. Here is one such example output:
$ gcc -std=c89 -Wall -Wextra -pedantic foo.c && ./a.out | wc -c 2147483649
It is worth noting that the above behaviour is only one of the many
possible ones. The code invokes undefined behaviour and the ISO
standard imposes no requirements on a specific implementation of the
compiler regarding what the behaviour of such code should be. For
example, an implementation could also exploit the undefined
behaviour to turn the loop into an infinite loop. In fact, GCC does
optimise it to an infinite loop if we compile the code with
the -O2
option.
# This never terminates! $ gcc -O2 -std=c89 -Wall -Wextra -pedantic foo.c && ./a.out
Let us take a look at the puzzle now.
Add or modify exactly one operator in the following code such that it prints exactly 6 dots.
for (i = 0; i < 6; i--) {
printf(".");
}
An obvious solution is to change i--
to i++
.
for (i = 0; i < 6; i++) {
printf(".");
}
There are a few more solutions to this puzzle. One of the solutions is very interesting. We will discuss the interesting solution in detail below.
Update on 02 Oct 2011: The puzzle has been solved in the comments section. We will discuss the solutions now. If you want to think about the problem before you see the solutions, this is a good time to pause and think about it. There are spoilers ahead.
Here is a list of some solutions:
for (i = 0; i < 6; i++)
for (i = 0; i < 6; ++i)
for (i = 0; -i < 6; i--)
for (i = 0; i + 6; i--)
for (i = 0; i ^= 6; i--)
The last solution involving the bitwise XOR operation is not immediately obvious. A little analysis is required to understand why it works.
Let us generalise the puzzle by replacing \( 6 \) in the loop with an arbitrary positive integer \( n. \) The loop in the last solution now becomes:
for (i = 0; i ^= n; i--) {
printf(".");
}
If we denote the value of the variable i
set by the
execution of i ^= n
after \( k \) dots are printed as
\( f(k), \) then
\[
f(k) =
\begin{cases}
0 & \text{if } k = 0, \\
n \oplus (f(k - 1) - 1) & \text{if } k > 1
\end{cases}
\]
where \( k \) is a nonnegative integer, \( n \) is a positive
integer, and the symbol \( \oplus \) denotes bitwise XOR operation
on two nonnegative integers.
Note that \( f(0) \) represents the value of i
set by
the execution of i ^= n
when no dots have been printed
yet.
If we can show that \( n \) is the least value of \( k \) for which \( f(k) = 0, \) it would prove that the loop terminates after printing \( n \) dots.
We will see in the next section that for odd values of \( n, \) \[ f(k) = \begin{cases} n & \text{if } k \text{ is even}, \\ 1 & \text{if } k \text{ is odd}. \end{cases} \] Therefore there is no value of \( k \) for which \( f(k) = 0 \) when \( n \) is odd. As a result, the loop never terminates when \( n \) is odd.
We will then see that for even values of \( n \) and \( 0 \leq k \leq n, \) \[ f(k) = 0 \iff k = n. \] Therefore the loop terminates after printing \( n \) dots when \( n \) is even.
We will first prove a few lemmas about some interesting properties of the bitwise XOR operation. We will then use it to prove the claims made in the previous section.
Lemma 1. For an odd positive integer \( n, \) \[ n \oplus (n - 1) = 1 \] where the symbol \( \oplus \) denotes bitwise XOR operation on two nonnegative integers.
Proof. Let the binary representation of \( n \) be \( b_m \dots b_1 b_0 \) where \( m \) is a nonnegative integer and \( b_m \) represents the most significant nonzero bit of \( n. \) Since \( n \) is an odd number, \( b_0 = 1. \) Thus \( n \) may be written as \[ b_m \dots b_1 1. \] As a result \( n - 1 \) may be written as \[ b_m \dots b_1 0. \] The bitwise XOR of both binary representations is \( 1. \)
Lemma 2. For a nonnegative integer \( n, \) \[ n \oplus 1 = \begin{cases} n + 1 & \text{if } n \text{ is even}, \\ n - 1 & \text{if } n \text{ is odd}. \end{cases} \] where the symbol \( \oplus \) denotes bitwise XOR operation on two nonnegative integers.
Proof. Let the binary representation of \( n \) be \( b_m \dots b_1 b_0 \) where \( m \) is a nonnegative integer and \( b_m \) represents the most significant nonzero bit of \( n. \)
If \( n \) is even, \( b_0 = 0. \) In this case, \( n \) may be written as \( b_m \dots b_1 0. \) Thus \( n \oplus 1 \) may be written as \( b_m \dots b_1 1. \) Therefore \( n \oplus 1 = n + 1. \)
If \( n \) is odd, \( b_0 = 1. \) In this case, \( n \) may be written as \( b_m \dots b_1 1. \) Thus \( n \oplus 1 \) may be written as \( b_m \dots b_1 0. \) Therefore \( n \oplus 1 = n - 1. \)
Note that for odd \( n, \) lemma 1 can also be derived as a corollary of lemma 2 in this manner: \[ k \oplus (k - 1) = k \oplus (k \oplus 1) = (k \oplus k) \oplus 1 = 0 \oplus 1 = 1. \]
Lemma 3. If \( x \) is an even nonnegative integer and \( y \) is an odd positive integer, then \( x \oplus y \) is odd, where the symbol \( \oplus \) denotes bitwise XOR operation on two nonnegative integers.
Proof. Let the binary representation of \( x \) be \( b_{xm_x} \dots b_{x1} b_{x0} \) and that of \( y \) be \( b_{ym_y} \dots b_{y1} b_{y0} \) where \( m_x \) and \( m_y \) are nonnegative integers and \( b_{xm_x} \) and \( b_{xm_y} \) represent the most significant nonzero bits of \( x \) and \( y, \) respectively.
Since \( x \) is even, \( b_{x0} = 0. \) Since \( y \) is odd, \( b_{y0} = 1. \)
Let \( z = x \oplus y \) with a binary representation of \( b_{zm_z} \dots b_{z1} b_{z0} \) where \( m_{zm_z} \) is a nonnegative integer and \( b_{zm_z} \) is the most significant nonzero bit of \( z. \)
We get \( b_{z0} = b_{x0} \oplus b_{y0} = 0 \oplus 1 = 1. \) Therefore \( z \) is odd.
Theorem 1. Let \( \oplus \) denote bitwise XOR operation on two nonnegative integers and \[ f(k) = \begin{cases} n & \text{if } n = 0, \\ n \oplus (f(n - 1) - 1) & \text{if } n > 1. \end{cases} \] where \( k \) is a nonnegative integer and \( n \) is an odd positive integer. Then \[ f(k) = \begin{cases} n & \text{if } k \text{ is even}, \\ 1 & \text{if } k \text{ is odd}. \end{cases} \]
Proof. This is a proof by mathematical induction. We have \( f(0) = n \) by definition. Therefore the base case holds good.
Let us assume that \( f(k) = n \) for any even \( k \) (induction hypothesis). Let \( k' = k + 1 \) and \( k'' = k + 2. \)
If \( k \) is even, we get \begin{align*} f(k') & = n \oplus (f(k) - 1) && \text{(by definition)} \\ & = n \oplus (n - 1) && \text{(by induction hypothesis)} \\ & = 1 && \text{(by lemma 1)},\\ f(k'') & = n \oplus (f(k') - 1) && \text{(by definition)} \\ & = n \oplus (1 - 1) && \text{(since \( f(k') = 1 \))} \\ & = n \oplus 0 \\ & = n. \end{align*}
Since \( f(k'') = n \) and \( k'' \) is the next even number after \( k, \) the induction step is complete. The induction step shows that for every even \( k, \) \( f(k) = n \) holds good. It also shows that as a result of \( f(k) = n \) for every even \( k, \) we get \( f(k') = 1 \) for every odd \( k'. \)
Theorem 2. Let \( \oplus \) denote bitwise XOR operation on two nonnegative integers and \[ f(k) = \begin{cases} n & \text{if } n = 0, \\ n \oplus (f(n - 1) - 1) & \text{if } n > 1. \end{cases} \] where \( k \) is a nonnegative integer, \( n \) is an even positive integer, and \( 0 \leq k \leq n. \) Then \[ f(k) = 0 \iff k = n. \]
Proof. We will first show by the principle of mathematical induction that for even \( k, \) \( f(k) = n - k. \) We have \( f(0) = n \) by definition, so the base case holds good. Now let us assume that \( f(k) = n - k \) holds good for any even \( k \) where \( 0 \leq k \leq n \) (induction hypothesis).
Since \( n \) is even (by definition) and \( k \) is even (by induction hypothesis), \( f(k) = n - k \) is even. As a result, \( f(k) - 1 \) is odd. By lemma 3, we conclude that \( f(k + 1) = n \oplus (f(k) - 1) \) is odd.
Now we perform the induction step as follows: \begin{align*} f(k + 2) & = n \oplus (f(k + 1) - 1) && \text{(by definition)} \\ & = n \oplus (f(k + 1) \oplus 1) && \text{(by lemma 2 for odd \( n \))} \\ & = n \oplus ((n \oplus (f(k) - 1)) \oplus 1) && \text{(by definition)} \\ & = (n \oplus n ) \oplus ((f(k) - 1) \oplus 1) && \text{(by associativity of XOR)} \\ & = 0 \oplus ((f(k) - 1) \oplus 1) \\ & = (f(k) - 1) \oplus 1 \\ & = (f(k) - 1) - 1 && \text{(from lemma 2 for odd \( n \))} \\ & = f(k) - 2 \\ & = n - k - 2 && \text{(by induction hypothesis).} \end{align*} This completes the induction step and proves that \( f(k) = n - k \) for even \( k \) where \( 0 \leq k \leq n. \)
We have shown above that \( f(k) \) is even for every even \( k \) where \( 0 \leq k \leq n \) which results in \( f(k + 1) \) as odd for every odd \( k + 1. \) This means that \( f(k) \) cannot be \( 0 \) for any odd \( k. \) Therefore \( f(k) = 0 \) is possible only even \( k. \) Solving \( f(k) = n - k = 0, \) we conclude that \( f(k) = 0 \) if and only if \( k = n. \)
Read on website | #c | #programming | #technology | #mathematics | #puzzle
]]>