Introduction to probability theory with an emphasis on the multivariate case. Includes random vectors, random processes, Markov chains, limit theorems, and related mathematics such as metric spaces, measure theory, and integration.
Presents a survey of the history and evolution of the branch of mathematics that focuses on probability and statistics, including useful applications and notable mathematicians in this area.
Unlike traditional introductory math/stat textbooks, Probability and Statistics: The Science of Uncertainty brings a modern flavor based on incorporating the computer to the course and an integrated approach to inference. From the start the book integrates simulations into its theoretical coverage, and emphasizes the use of computer-powered computation throughout.* Math and science majors with just one year of calculus can use this text and experience a refreshing blend of applications and theory that goes beyond merely mastering the technicalities. They'll get a thorough grounding in probability theory, and go beyond that to the theory of statistical inference and its applications. An integrated approach to inference is presented that includes the frequency approach as well as Bayesian methodology. Bayesian inference is developed as a logical extension of likelihood methods. A separate chapter is devoted to the important topic of model checking and this is applied in the context of the standard applied statistical techniques. Examples of data analyses using real-world data are presented throughout the text. A final chapter introduces a number of the most important stochastic process models using elementary methods. *Note: An appendix in the book contains Minitab code for more involved computations. The code can be used by students as templates for their own calculations. If a software package like Minitab is used with the course then no programming is required by the students.
Together with the fundamentals of probability, random processes and statistical analysis, this insightful book also presents a broad range of advanced topics and applications. There is extensive coverage of Bayesian vs. frequentist statistics, time series and spectral representation, inequalities, bound and approximation, maximum-likelihood estimation and the expectation-maximization (EM) algorithm, geometric Brownian motion and Itô process. Applications such as hidden Markov models (HMM), the Viterbi, BCJR, and Baum–Welch algorithms, algorithms for machine learning, Wiener and Kalman filters, and queueing and loss networks are treated in detail. The book will be useful to students and researchers in such areas as communications, signal processing, networks, machine learning, bioinformatics, econometrics and mathematical finance. With a solutions manual, lecture slides, supplementary materials and MATLAB programs all available online, it is ideal for classroom teaching as well as a valuable reference for professionals.
A comprehensive and up-to-date introduction to the mathematics that all economics students need to know Probability theory is the quantitative language used to handle uncertainty and is the foundation of modern statistics. Probability and Statistics for Economists provides graduate and PhD students with an essential introduction to mathematical probability and statistical theory, which are the basis of the methods used in econometrics. This incisive textbook teaches fundamental concepts, emphasizes modern, real-world applications, and gives students an intuitive understanding of the mathematics that every economist needs to know. Covers probability and statistics with mathematical rigor while emphasizing intuitive explanations that are accessible to economics students of all backgrounds Discusses random variables, parametric and multivariate distributions, sampling, the law of large numbers, central limit theory, maximum likelihood estimation, numerical optimization, hypothesis testing, and more Features hundreds of exercises that enable students to learn by doing Includes an in-depth appendix summarizing important mathematical results as well as a wealth of real-world examples Can serve as a core textbook for a first-semester PhD course in econometrics and as a companion book to Bruce E. Hansen’s Econometrics Also an invaluable reference for researchers and practitioners
Using only the very elementary framework of finite probability spaces, this book treats a number of topics in the modern theory of stochastic processes. This is made possible by using a small amount of Abraham Robinson's nonstandard analysis and not attempting to convert the results into conventional form.
Classical probability theory and mathematical statistics appear sometimes too rigid for real life problems, especially while dealing with vague data or imprecise requirements. These problems have motivated many researchers to "soften" the classical theory. Some "softening" approaches utilize concepts and techniques developed in theories such as fuzzy sets theory, rough sets, possibility theory, theory of belief functions and imprecise probabilities, etc. Since interesting mathematical models and methods have been proposed in the frameworks of various theories, this text brings together experts representing different approaches used in soft probability, statistics and data analysis.
This text presents notions and ideas at the foundations of a statistical treatment of risks. The focus is on statistical applications within the field of engineering risk and safety analysis. Coverage includes Bayesian methods. Such knowledge facilitates the understanding of the influence of random phenomena and gives a deeper understanding of the role of probability in risk analysis. The text is written for students who have studied elementary undergraduate courses in engineering mathematics, perhaps including a minor course in statistics. This book differs from typical textbooks in its verbal approach to many explanations and examples.
This book is a fresh approach to a calculus based, first course in probability and statistics, using R throughout to give a central role to data and simulation. The book introduces probability with Monte Carlo simulation as an essential tool. Simulation makes challenging probability questions quickly accessible and easily understandable. Mathematical approaches are included, using calculus when appropriate, but are always connected to experimental computations. Using R and simulation gives a nuanced understanding of statistical inference. The impact of departure from assumptions in statistical tests is emphasized, quantified using simulations, and demonstrated with real data. The book compares parametric and non-parametric methods through simulation, allowing for a thorough investigation of testing error and power. The text builds R skills from the outset, allowing modern methods of resampling and cross validation to be introduced along with traditional statistical techniques. Fifty-two data sets are included in the complementary R package fosdata. Most of these data sets are from recently published papers, so that you are working with current, real data, which is often large and messy. Two central chapters use powerful tidyverse tools (dplyr, ggplot2, tidyr, stringr) to wrangle data and produce meaningful visualizations. Preliminary versions of the book have been used for five semesters at Saint Louis University, and the majority of the more than 400 exercises have been classroom tested.
An up-to-date, comprehensive treatment of a classic text on missing data in statistics The topic of missing data has gained considerable attention in recent decades. This new edition by two acknowledged experts on the subject offers an up-to-date account of practical methodology for handling missing data problems. Blending theory and application, authors Roderick Little and Donald Rubin review historical approaches to the subject and describe simple methods for multivariate analysis with missing values. They then provide a coherent theory for analysis of problems based on likelihoods derived from statistical models for the data and the missing data mechanism, and then they apply the theory to a wide range of important missing data problems. Statistical Analysis with Missing Data, Third Edition starts by introducing readers to the subject and approaches toward solving it. It looks at the patterns and mechanisms that create the missing data, as well as a taxonomy of missing data. It then goes on to examine missing data in experiments, before discussing complete-case and available-case analysis, including weighting methods. The new edition expands its coverage to include recent work on topics such as nonresponse in sample surveys, causal inference, diagnostic methods, and sensitivity analysis, among a host of other topics. An updated “classic” written by renowned authorities on the subject Features over 150 exercises (including many new ones) Covers recent work on important methods like multiple imputation, robust alternatives to weighting, and Bayesian methods Revises previous topics based on past student feedback and class experience Contains an updated and expanded bibliography The authors were awarded The Karl Pearson Prize in 2017 by the International Statistical Institute, for a research contribution that has had profound influence on statistical theory, methodology or applications. Their work "has been no less than defining and transforming." (ISI) Statistical Analysis with Missing Data, Third Edition is an ideal textbook for upper undergraduate and/or beginning graduate level students of the subject. It is also an excellent source of information for applied statisticians and practitioners in government and industry.