Digitization of DeGroot and Schervish, Probability and Statistics (Fourth Edition), by Jeff Jacobs, Assistant Teaching Professor of Data Science and Analytics, Georgetown University.
Changes in the Digitized Version¶
To match the notation we’re using in DSAN 5100, I have made the following changes (which could be undone via a find-and-replace, if you want to change the format back to the original textbook format):
The book uses the term probability function, and the abbreviation p.f., to refer to the function that gives us the probability that a discrete random variable takes on the value . To match the notation we use in class, I have changed this throughout to use the term probability mass function, and the abbreviation pmf.
Similarly, I make sure to use (lowercased) pdf to refer to the probability density function, and CDF (uppercased) to refer to the cumulative density function.
Changes to the Fourth Edition¶
I have reorganized many main results that were included in the body of the text by labeling them as theorems in order to facilitate students in finding and referencing these results.
I have pulled the important defintions and assumptions out of the body of the text and labeled them as such so that they stand out better.
When a new topic is introduced, I introduce it with a motivating example before delving into the mathematical formalities. Then I return to the example to illustrate the newly introduced material.
I moved the material on the law of large numbers and the central limit theorem to a new sec-6. It seemed more natural to deal with the main large-sample results together.
I moved the section on Markov chains into Chapter 3: Random Variables and Distributions. Every time I cover this material with my own students, I stumble over not being able to refer to random variables, distributions, and conditional distributions. I have actually postponed this material until after introducing distributions, and then gone back to cover Markov chains. I feel that the time has come to place it in a more natural location. I also added some material on stationary distributions of Markov chains.
I have moved the lengthy proofs of several theorems to the ends of their respective sections in order to improve the flow of the presentation of ideas.
I rewrote sec-7-1 to make the introduction to inference clearer.
I rewrote sec-9-1 as a more complete introduction to hypothesis testing, including likelihood ratio tests. For instructors not interested in the more mathematical theory of hypothesis testing, it should now be easier to skip from sec-9-1 directly to sec-9-5.
Some other changes that readers will notice:
I have replaced the notation in which the intersection of two sets and had been represented with the more popular . The old notation, although mathematically sound, seemed a bit arcane for a text at this level.
I added the statements of Stirling’s formula and Jensen’s inequality.
I moved the law of total probability and the discussion of partitions of a sample space from 2.3 Bayes’ Theorem to 2.1 The Definition of Conditional Probability.
I define the cumulative distribution function (CDF) as the prefered name of what used to be called only the distribution function (DF).
I added some discussion of histograms in Chapter 3: Random Variables and Distributions and sec-6.
I rearranged the topics in 3.8 Functions of a Random Variable and 3.9 Functions of Two or More Random Variables so that simple functions of random variables appear first and the general formulations appear at the end to make it easier for instructors who want to avoid some of the more mathematically challenging parts.
I emphasized the closeness of a hypergeometric distribution with a large number of available items to a binomial distribution.
I gave a brief introduction to Chernoff bounds. These are becoming increasingly important in computer science, and their derivation requires only material that is already in the text.
I changed the definition of confidence interval to refer to the random interval rather than the observed interval. This makes statements less cumbersome, and it corresponds to more modern usage.
I added a brief discussion of the method of moments in sec-7-6.
I added brief introductions to Newton’s method and the EM algorithm in sec-7.
I introduced the concept of pivotal quantity to facilitate construction of confidence intervals in general.
I added the statement of the large-sample distribution of the likelihood ratio test statistic. I then used this as an alternative way to test the null hypothesis that two normal means are equal when it is not assumed that the variances are equal.
I moved the Bonferroni inequality into the main text (Chapter 1: Introduction to Probability) and later (sec-11) used it as a way to construct simultaneous tests and confidence intervals.
How to Use This Book¶
The text is somewhat long for complete coverage in a one-year course at the undergraduate level and is designed so that instructors can make choices about which topics are most important to cover and which can be left for more in-depth study. As an example, many instructors wish to deemphasize the classical counting arguments that are detailed in Sections 1.7 Counting Methods–1.9 Multinomial Coefficients. An instructor who only wants enough information to be able to cover the binomial and/or multinomial distributions can safely discuss only the definitions and theorems on permutations, combinations, and possibly multinomial coefficients. Just make sure that the students realize what these values count, otherwise the associated distributions will make no sense. The various examples in these sections are helpful, but not necessary, for understanding the important distributions. Another example is 3.9 Functions of Two or More Random Variables on functions of two or more random variables. The use of Jacobians for general multivariate transformations might be more mathematics than the instructors of some undergraduate courses are willing to cover. The entire section could be skipped without causing problems later in the course, but some of the more straightforward cases early in the section (such as convolution) might be worth introducing. The material in Sections sec-9-2–sec-9-4 on optimal tests in one-parameter families is pretty mathematics, but it is of interest primarily to graduate students who require a very deep understanding of hypothesis testing theory. The rest of sec-9 covers everything that an undergraduate course really needs.
In addition to the text, the publisher has an Instructor’s Solutions Manual, available for download from the Instructor Resource Center at www
I have taught a year-long probability and statistics sequence from earlier editions of this text for a group of mathematically well-trained juniors and seniors. In the first semester, I covered what was in the earlier edition but is now in the first five chapters (including the material on Markov chains) and parts of sec-6. In the second semester, I covered the rest of the new sec-6, Chapters sec-7–sec-9, Sections sec-11-1-sec-11-5, and sec-12. I have also taught a one-semester probability and random processes course for engineers and computer scientists. I covered what was in the old edition and is now in Chapters Chapter 1: Introduction to Probability–sec-6 and sec-12, including Markov chains, but not Jacobians. This latter course did not emphasize mathematical derivation to the same extent as the course for mathematics students.
A number of sections are designated with an asterisk (*). This indicates that later sections do not rely materially on the material in that section. This designation is not intended to suggest that instructors skip these sections. Skipping one of these sections will not cause the students to miss definitions or results that they will need later. The sections are 2.4 The Gambler’s Ruin Problem, 3.10 Markov Chains, 4.8 Utility, sec-7-7, sec-7-8, sec-7-9, sec-8-6, sec-8-8, sec-9-2, sec-9-3, sec-9-4, sec-9-8, sec-9-9, sec-10-6, sec-10-7, sec-10-8, sec-11-4, sec-11-7, sec-11-8, and sec-12-5. Aside from cross-references between sections within this list, occasional material from elsewhere in the text does refer back to some of the sections in this list. Each of the dependencies is quite minor, however.
Most of the dependencies involve references from sec-12 back to one of the optional sections. The reason for this is that the optional sections address some of the more difficult material, and simulation is most useful for solving those difficult problems that cannot be solved analytically. Except for passing references that help put material into context, the dependencies are as follows:
The sample distribution function (sec-10-6) is reintroduced during the discussion of the bootstrap in sec-12-6. The sample distribution function is also a useful tool for displaying simulation results. It could be introduced as early as exm-12-3-7 simply by covering the first subsection of sec-10-6.
The material on robust estimation (sec-10-7) is revisited in some simulation exercises in sec-12-2 (Exercises exr-12-2-4, exr-12-2-5, exr-12-2-7, and exr-12-2-8).
exm-12-3-4 makes reference to the material on two-way analysis of variance (Sections sec-11-7 and sec-11-8).
Supplements¶
The text is accompanied by the following supplementary material:
Instructor’s Solutions Manual contains fully worked solutions to all exercises in the text. Available for download from the Instructor Resource Center at www
.pearsonhighered .com /irc. Student Solutions Manual contains fully worked solutions to all odd exercises in the text. Available for purchase from MyPearsonStore at www
.mypearsonstore .com. (ISBN-13: 978-0-321-71598-2; ISBN-10: 0-321-71598-5)
Andre Adler, Illinois Institute of Technology
E.N. Barron, Loyola University
Brian Blank,Washington University in St. Louis
Indranil Chakraborty, University of Oklahoma
Daniel Chambers, Boston College
Rita Chattopadhyay, Eastern Michigan University
Stephen A. Chiappari, Santa Clara University
Sheng-Kai Chang, Wayne State University
Justin Corvino, Lafayette College
Michael Evans, University of Toronto
Doug Frank, Indiana University of Pennsylvania
Anda Gadidov, Kennesaw State University
Lyn Geisler, Randolph–Macon College
Prem Goel, Ohio State University
Susan Herring, Sonoma State University
Pawel Hitczenko, Drexel University
Lifang Hsu, Le Moyne College
Wei-Min Huang, Lehigh University
Syed Kirmani, University of Northern Iowa
Michael Lavine, Duke University
Rich Levine, San Diego State University
John Liukkonen, Tulane University
Sergio Loch, Grand View College
Rosa Matzkin, Northwestern University
Terry McConnell, Syracuse University
Hans-Georg Mueller, University of California–Davis
Robert Myers, Bethel College
Mario Peruggia, The Ohio State University
Stefan Ralescu, Queens University
Krishnamurthi Ravishankar, SUNY New Paltz
Diane Saphire, Trinity University
Steven Sepanski, Saginaw Valley State University
Hen-Siong Tan, Pennsylvania University
Kanapathi Thiru, University of Alaska
Kenneth Troske, Johns Hopkins University
John Van Ness, University of Texas at Dallas
Yehuda Vardi, Rutgers University
Yelena Vaynberg, Wayne State University
Joseph Verducci, Ohio State University
Mahbobeh Vezveai, Kent State University
Brani Vidakovic, Duke University
Karin Vorwerk, Westfield State College
Bette Warren, Eastern Michigan University
Calvin L. Williams, Clemson University
Lori Wolff, University of Mississippi
The person who checked the accuracy of the book was Anda Gadidov, Kennesaw State University. I would also like to thank my colleagues at Carnegie Mellon University, especially Anthony Brockwell, Joel Greenhouse, John Lehoczky, Heidi Sestrich, and Valerie Ventura.
The people at Addison-Wesley and other organizations that helped produce the book were Paul Anagnostopoulos, Patty Bergin, Dana Jones Bettez, Chris Cummings, Kathleen DeChavez, Alex Gay, Leah Goldberg, Karen Hartpence, and Christina Lepre.
If I left anyone out, it was unintentional, and I apologize. Errors inevitably arise in any project like this (meaning a project in which I am involved). For this reason, I shall post information about the book, including a list of corrections, on my Web page, https://
Mark J. Schervish
October 2020
Acknowledgments¶
There are many people that I want to thank for their help and encouragement during this revision. First and foremost, I want to thank Marilyn DeGroot and Morrie’s children for giving me the chance to revise Morrie’s masterpiece.
I am indebted to the many readers, reviewers, colleagues, staff, and people at Addison-Wesley whose help and comments have strengthened this edition. The reviewers were: