Written by one of the main figures in twentieth century statistics, this book provides a unified treatment of first-order large-sample theory. It discusses a broad range of applications including introductions to density estimation, the bootstrap, and the asymptotics of survey methodology. The book is written at an elementary level making it accessible to most readers.
A Course in Large Sample Theory is presented in four parts. The first treats basic probabilistic notions, the second features the basic statistical tools for expanding the theory, the third contains special topics as applications of the general theory, and the fourth covers more standard statistical topics. Nearly all topics are covered in their multivariate setting.The book is intended as a first year graduate course in large sample theory for statisticians. It has been used by graduate students in statistics, biostatistics, mathematics, and related fields. Throughout the book there are many examples and exercises with solutions. It is an ideal text for self study.
In a way, the world is made up of approximations, and surely there is no exception in the world of statistics. In fact, approximations, especially large sample approximations, are very important parts of both theoretical and - plied statistics.TheGaussiandistribution,alsoknownasthe normaldistri- tion,is merelyonesuchexample,dueto thewell-knowncentrallimittheorem. Large-sample techniques provide solutions to many practical problems; they simplify our solutions to di?cult, sometimes intractable problems; they j- tify our solutions; and they guide us to directions of improvements. On the other hand, just because large-sample approximations are used everywhere, and every day, it does not guarantee that they are used properly, and, when the techniques are misused, there may be serious consequences. 2 Example 1 (Asymptotic? distribution). Likelihood ratio test (LRT) is one of the fundamental techniques in statistics. It is well known that, in the 2 “standard” situation, the asymptotic null distribution of the LRT is?,with the degreesoffreedomequaltothe di?erencebetweenthedimensions,de?ned as the numbers of free parameters, of the two nested models being compared (e.g., Rice 1995, pp. 310). This might lead to a wrong impression that the 2 asymptotic (null) distribution of the LRT is always? . A similar mistake 2 might take place when dealing with Pearson’s? -test—the asymptotic distri- 2 2 bution of Pearson’s? -test is not always? (e.g., Moore 1978).
This unique book delivers an encyclopedic treatment of classic as well as contemporary large sample theory, dealing with both statistical problems and probabilistic issues and tools. The book is unique in its detailed coverage of fundamental topics. It is written in an extremely lucid style, with an emphasis on the conceptual discussion of the importance of a problem and the impact and relevance of the theorems. There is no other book in large sample theory that matches this book in coverage, exercises and examples, bibliography, and lucid conceptual discussion of issues and theorems.
Elements of Probability Theory focuses on the basic ideas and methods of the theory of probability. The book first discusses events and probabilities, including the classical meaning of probability, fundamental properties of probabilities, and the primary rule for the multiplication of probabilities. The text also touches on random variables and probability distributions. Topics include discrete and random variables; functions of random variables; and binomial distributions. The selection also discusses the numerical characteristics of probability distributions; limit theorems and estimates of the mean; and the law of large numbers. The text also describes linear correlation, including conditional expectations and their properties, coefficient of correlation, and best linear approximation to the regression function. The book presents tables that show the values of the normal probability integral, Poisson distribution, and values of the normal probability density. The text is a good source of data for readers and students interested in probability theory.
High-dimensional data appear in many fields, and their analysis has become increasingly important in modern statistics. However, it has long been observed that several well-known methods in multivariate analysis become inefficient, or even misleading, when the data dimension p is larger than, say, several tens. A seminal example is the well-known inefficiency of Hotelling's T2-test in such cases. This example shows that classical large sample limits may no longer hold for high-dimensional data; statisticians must seek new limiting theorems in these instances. Thus, the theory of random matrices (RMT) serves as a much-needed and welcome alternative framework. Based on the authors' own research, this book provides a first-hand introduction to new high-dimensional statistical methods derived from RMT. The book begins with a detailed introduction to useful tools from RMT, and then presents a series of high-dimensional problems with solutions provided by RMT methods.
Introductory Business Statistics 2e aligns with the topics and objectives of the typical one-semester statistics course for business, economics, and related majors. The text provides detailed and supportive explanations and extensive step-by-step walkthroughs. The author places a significant emphasis on the development and practical application of formulas so that students have a deeper understanding of their interpretation and application of data. Problems and exercises are largely centered on business topics, though other applications are provided in order to increase relevance and showcase the critical role of statistics in a number of fields and real-world contexts. The second edition retains the organization of the original text. Based on extensive feedback from adopters and students, the revision focused on improving currency and relevance, particularly in examples and problems. This is an adaptation of Introductory Business Statistics 2e by OpenStax. You can access the textbook as pdf for free at openstax.org. Minor editorial changes were made to ensure a better ebook reading experience. Textbook content produced by OpenStax is licensed under a Creative Commons Attribution 4.0 International License.
Taken literally, the title "All of Statistics" is an exaggeration. But in spirit, the title is apt, as the book does cover a much broader range of topics than a typical introductory book on mathematical statistics. This book is for people who want to learn probability and statistics quickly. It is suitable for graduate or advanced undergraduate students in computer science, mathematics, statistics, and related disciplines. The book includes modern topics like non-parametric curve estimation, bootstrapping, and classification, topics that are usually relegated to follow-up courses. The reader is presumed to know calculus and a little linear algebra. No previous knowledge of probability and statistics is required. Statistics, data mining, and machine learning are all concerned with collecting and analysing data.
The latest edition of this classic is updated with new problem sets and material The Second Edition of this fundamental textbook maintains the book's tradition of clear, thought-provoking instruction. Readers are provided once again with an instructive mix of mathematics, physics, statistics, and information theory. All the essential topics in information theory are covered in detail, including entropy, data compression, channel capacity, rate distortion, network information theory, and hypothesis testing. The authors provide readers with a solid understanding of the underlying theory and applications. Problem sets and a telegraphic summary at the end of each chapter further assist readers. The historical notes that follow each chapter recap the main points. The Second Edition features: * Chapters reorganized to improve teaching * 200 new problems * New material on source coding, portfolio theory, and feedback capacity * Updated references Now current and enhanced, the Second Edition of Elements of Information Theory remains the ideal textbook for upper-level undergraduate and graduate courses in electrical engineering, statistics, and telecommunications.
Intended as the text for a sequence of advanced courses, this book covers major topics in theoretical statistics in a concise and rigorous fashion. The discussion assumes a background in advanced calculus, linear algebra, probability, and some analysis and topology. Measure theory is used, but the notation and basic results needed are presented in an initial chapter on probability, so prior knowledge of these topics is not essential. The presentation is designed to expose students to as many of the central ideas and topics in the discipline as possible, balancing various approaches to inference as well as exact, numerical, and large sample methods. Moving beyond more standard material, the book includes chapters introducing bootstrap methods, nonparametric regression, equivariant estimation, empirical Bayes, and sequential design and analysis. The book has a rich collection of exercises. Several of them illustrate how the theory developed in the book may be used in various applications. Solutions to many of the exercises are included in an appendix.