This text presents introductory statistics from the perspective of real-world situations. Each section within a chapter starts with a realistic data setting illustrating the corresponding method. Throughout the text, students are encouraged to use graphic displays to get a feeling for the information contained in the sample data and to check the reasonableness of the corresponding assumptions.
Statistics and hypothesis testing are routinely used in areas (such as linguistics) that are traditionally not mathematically intensive. In such fields, when faced with experimental data, many students and researchers tend to rely on commercial packages to carry out statistical data analysis, often without understanding the logic of the statistical tests they rely on. As a consequence, results are often misinterpreted, and users have difficulty in flexibly applying techniques relevant to their own research — they use whatever they happen to have learned. A simple solution is to teach the fundamental ideas of statistical hypothesis testing without using too much mathematics. This book provides a non-mathematical, simulation-based introduction to basic statistical concepts and encourages readers to try out the simulations themselves using the source code and data provided (the freely available programming language R is used throughout). Since the code presented in the text almost always requires the use of previously introduced programming constructs, diligent students also acquire basic programming abilities in R. The book is intended for advanced undergraduate and graduate students in any discipline, although the focus is on linguistics, psychology, and cognitive science. It is designed for self-instruction, but it can also be used as a textbook for a first course on statistics. Earlier versions of the book have been used in undergraduate and graduate courses in Europe and the US. ”Vasishth and Broe have written an attractive introduction to the foundations of statistics. It is concise, surprisingly comprehensive, self-contained and yet quite accessible. Highly recommended.” Harald Baayen, Professor of Linguistics, University of Alberta, Canada ”By using the text students not only learn to do the specific things outlined in the book, they also gain a skill set that empowers them to explore new areas that lie beyond the book’s coverage.” Colin Phillips, Professor of Linguistics, University of Maryland, USA
Every topic in this book is introduced through the use of real-world applications, that use real data. Each section starts with realistic data that illustrates the need for the method introduced. The data and other examples are then used to demonstrate the methods. Throughout the book, readers are encouraged to use graphical displays to get a feeling for the information contained in the sample data.
This broad text provides a complete overview of most standard statistical methods, including multiple regression, analysis of variance, experimental design, and sampling techniques. Assuming a background of only two years of high school algebra, this book teaches intelligent data analysis and covers the principles of good data collection. * Provides a complete discussion of analysis of data including estimation, diagnostics, and remedial actions * Examples contain graphical illustration for ease of interpretation * Intended for use with almost any statistical software * Examples are worked to a logical conclusion, including interpretation of results * A complete Instructor's Manual is available to adopters
This book is a fresh approach to a calculus based, first course in probability and statistics, using R throughout to give a central role to data and simulation. The book introduces probability with Monte Carlo simulation as an essential tool. Simulation makes challenging probability questions quickly accessible and easily understandable. Mathematical approaches are included, using calculus when appropriate, but are always connected to experimental computations. Using R and simulation gives a nuanced understanding of statistical inference. The impact of departure from assumptions in statistical tests is emphasized, quantified using simulations, and demonstrated with real data. The book compares parametric and non-parametric methods through simulation, allowing for a thorough investigation of testing error and power. The text builds R skills from the outset, allowing modern methods of resampling and cross validation to be introduced along with traditional statistical techniques. Fifty-two data sets are included in the complementary R package fosdata. Most of these data sets are from recently published papers, so that you are working with current, real data, which is often large and messy. Two central chapters use powerful tidyverse tools (dplyr, ggplot2, tidyr, stringr) to wrangle data and produce meaningful visualizations. Preliminary versions of the book have been used for five semesters at Saint Louis University, and the majority of the more than 400 exercises have been classroom tested.
The chief executive officer of a corporation is not much different from a public school administrator. While CEOs base many of their decisions on data, for school administrators, this type of research may conjure up miserable memories of searching for information to meet a graduate school requirement. However, the value of data-based decision making will continue to escalate and the school community—students, teachers, parents and the general public—expect this information to come from their administrators. Administrators are called on to be accountable, but few are capable of presenting the mountain of data that they collect in a cohesive and strategic manner. Most statistical books are focused on statistical theory versus application, but Statistics Made Simple for School Leaders presents statistics in a simple, practical, conceptual, and immediately applicable manner. It enables administrators to take their data and manage it into strategic information so the results can be used for action plans that benefit the school system. The approach is 'user friendly' and leaves the reader with a confident can-do attitude to communicate results and plans to staff and the community.
The environment for obtaining information and providing statistical data for policy makers and the public has changed significantly in the past decade, raising questions about the fundamental survey paradigm that underlies federal statistics. New data sources provide opportunities to develop a new paradigm that can improve timeliness, geographic or subpopulation detail, and statistical efficiency. It also has the potential to reduce the costs of producing federal statistics. The panel's first report described federal statistical agencies' current paradigm, which relies heavily on sample surveys for producing national statistics, and challenges agencies are facing; the legal frameworks and mechanisms for protecting the privacy and confidentiality of statistical data and for providing researchers access to data, and challenges to those frameworks and mechanisms; and statistical agencies access to alternative sources of data. The panel recommended a new approach for federal statistical programs that would combine diverse data sources from government and private sector sources and the creation of a new entity that would provide the foundational elements needed for this new approach, including legal authority to access data and protect privacy. This second of the panel's two reports builds on the analysis, conclusions, and recommendations in the first one. This report assesses alternative methods for implementing a new approach that would combine diverse data sources from government and private sector sources, including describing statistical models for combining data from multiple sources; examining statistical and computer science approaches that foster privacy protections; evaluating frameworks for assessing the quality and utility of alternative data sources; and various models for implementing the recommended new entity. Together, the two reports offer ideas and recommendations to help federal statistical agencies examine and evaluate data from alternative sources and then combine them as appropriate to provide the country with more timely, actionable, and useful information for policy makers, businesses, and individuals.
This book provides a comprehensive and up to date treatment of theory and practical implementation in Register-based statistics. It begins by defining the area, before explaining how to structure such systems, as well as detailing alternative approaches. It explains how to create statistical registers, how to implement quality assurance, and the use of IT systems for register-based statistics. Further to this, clear details are given about the practicalities of implementing such statistical methods, such as protection of privacy and the coordination and coherence of such an undertaking. This edition offers a full understanding of both the principles and practices of this increasingly popular area of statistics, and can be considered a first step to a more systematic way of working with register-statistical issues. This book addresses the growing global interest in the topic and employs a much broader, more international approach than the 1st edition. New chapters explore different kinds of register-based surveys, such as preconditions for register-based statistics and comparing sample survey and administrative data. Furthermore, the authors present discussions on register-based census, national accounts and the transition towards a register-based system as well as presenting new chapters on quality assessment of administrative sources and production process quality.
Statistical Analysis of Questionnaires: A Unified Approach Based on R and Stata presents special statistical methods for analyzing data collected by questionnaires. The book takes an applied approach to testing and measurement tasks, mirroring the growing use of statistical methods and software in education, psychology, sociology, and other fields.