Provides a simple exposition of the basic time series material, and insights into underlying technical aspects and methods of proof Long memory time series are characterized by a strong dependence between distant events. This book introduces readers to the theory and foundations of univariate time series analysis with a focus on long memory and fractional integration, which are embedded into the general framework. It presents the general theory of time series, including some issues that are not treated in other books on time series, such as ergodicity, persistence versus memory, asymptotic properties of the periodogram, and Whittle estimation. Further chapters address the general functional central limit theory, parametric and semiparametric estimation of the long memory parameter, and locally optimal tests. Intuitive and easy to read, Time Series Analysis with Long Memory in View offers chapters that cover: Stationary Processes; Moving Averages and Linear Processes; Frequency Domain Analysis; Differencing and Integration; Fractionally Integrated Processes; Sample Means; Parametric Estimators; Semiparametric Estimators; and Testing. It also discusses further topics. This book: Offers beginning-of-chapter examples as well as end-of-chapter technical arguments and proofs Contains many new results on long memory processes which have not appeared in previous and existing textbooks Takes a basic mathematics (Calculus) approach to the topic of time series analysis with long memory Contains 25 illustrative figures as well as lists of notations and acronyms Time Series Analysis with Long Memory in View is an ideal text for first year PhD students, researchers, and practitioners in statistics, econometrics, and any application area that uses time series over a long period. It would also benefit researchers, undergraduates, and practitioners in those areas who require a rigorous introduction to time series analysis.
Long-memory processes are known to play an important part in many areas of science and technology, including physics, geophysics, hydrology, telecommunications, economics, finance, climatology, and network engineering. In the last 20 years enormous progress has been made in understanding the probabilistic foundations and statistical principles of such processes. This book provides a timely and comprehensive review, including a thorough discussion of mathematical and probabilistic foundations and statistical methods, emphasizing their practical motivation and mathematical justification. Proofs of the main theorems are provided and data examples illustrate practical aspects. This book will be a valuable resource for researchers and graduate students in statistics, mathematics, econometrics and other quantitative areas, as well as for practitioners and applied researchers who need to analyze data in which long memory, power laws, self-similar scaling or fractal properties are relevant.
The field of financial econometrics has exploded over the last decade This book represents an integration of theory, methods, and examples using the S-PLUS statistical modeling language and the S+FinMetrics module to facilitate the practice of financial econometrics. This is the first book to show the power of S-PLUS for the analysis of time series data. It is written for researchers and practitioners in the finance industry, academic researchers in economics and finance, and advanced MBA and graduate students in economics and finance. Readers are assumed to have a basic knowledge of S-PLUS and a solid grounding in basic statistics and time series concepts. This Second Edition is updated to cover S+FinMetrics 2.0 and includes new chapters on copulas, nonlinear regime switching models, continuous-time financial models, generalized method of moments, semi-nonparametric conditional density models, and the efficient method of moments. Eric Zivot is an associate professor and Gary Waterman Distinguished Scholar in the Economics Department, and adjunct associate professor of finance in the Business School at the University of Washington. He regularly teaches courses on econometric theory, financial econometrics and time series econometrics, and is the recipient of the Henry T. Buechel Award for Outstanding Teaching. He is an associate editor of Studies in Nonlinear Dynamics and Econometrics. He has published papers in the leading econometrics journals, including Econometrica, Econometric Theory, the Journal of Business and Economic Statistics, Journal of Econometrics, and the Review of Economics and Statistics. Jiahui Wang is an employee of Ronin Capital LLC. He received a Ph.D. in Economics from the University of Washington in 1997. He has published in leading econometrics journals such as Econometrica and Journal of Business and Economic Statistics, and is the Principal Investigator of National Science Foundation SBIR grants. In 2002 Dr. Wang was selected as one of the "2000 Outstanding Scholars of the 21st Century" by International Biographical Centre.
Statistical Methods for Long Term Memory Processes covers the diverse statistical methods and applications for data with long-range dependence. Presenting material that previously appeared only in journals, the author provides a concise and effective overview of probabilistic foundations, statistical methods, and applications. The material emphasizes basic principles and practical applications and provides an integrated perspective of both theory and practice. This book explores data sets from a wide range of disciplines, such as hydrology, climatology, telecommunications engineering, and high-precision physical measurement. The data sets are conveniently compiled in the index, and this allows readers to view statistical approaches in a practical context. Statistical Methods for Long Term Memory Processes also supplies S-PLUS programs for the major methods discussed. This feature allows the practitioner to apply long memory processes in daily data analysis. For newcomers to the area, the first three chapters provide the basic knowledge necessary for understanding the remainder of the material. To promote selective reading, the author presents the chapters independently. Combining essential methodologies with real-life applications, this outstanding volume is and indispensable reference for statisticians and scientists who analyze data with long-range dependence.
A self-contained, contemporary treatment of the analysis of long-range dependent data Long-Memory Time Series: Theory and Methods provides an overview of the theory and methods developed to deal with long-range dependent data and describes the applications of these methodologies to real-life time series. Systematically organized, it begins with the foundational essentials, proceeds to the analysis of methodological aspects (Estimation Methods, Asymptotic Theory, Heteroskedastic Models, Transformations, Bayesian Methods, and Prediction), and then extends these techniques to more complex data structures. To facilitate understanding, the book: Assumes a basic knowledge of calculus and linear algebra and explains the more advanced statistical and mathematical concepts Features numerous examples that accelerate understanding and illustrate various consequences of the theoretical results Proves all theoretical results (theorems, lemmas, corollaries, etc.) or refers readers to resources with further demonstration Includes detailed analyses of computational aspects related to the implementation of the methodologies described, including algorithm efficiency, arithmetic complexity, CPU times, and more Includes proposed problems at the end of each chapter to help readers solidify their understanding and practice their skills A valuable real-world reference for researchers and practitioners in time series analysis, economerics, finance, and related fields, this book is also excellent for a beginning graduate-level course in long-memory processes or as a supplemental textbook for those studying advanced statistics, mathematics, economics, finance, engineering, or physics. A companion Web site is available for readers to access the S-Plus and R data sets used within the text.
Assembles three different strands of long memory analysis: statistical literature on the properties of, and tests for, LRD processes; mathematical literature on the stochastic processes involved; and models from economic theory providing plausible micro foundations for the occurrence of long memory in economics.
Practice makes perfect. Therefore the best method of mastering models is working with them. This book contains a large collection of exercises and solutions which will help explain the statistics of financial markets. These practical examples are carefully presented and provide computational solutions to specific problems, all of which are calculated using R and Matlab. This study additionally looks at the concept of corresponding Quantlets, the name given to these program codes and which follow the name scheme SFSxyz123. The book is divided into three main parts, in which option pricing, time series analysis and advanced quantitative statistical techniques in finance is thoroughly discussed. The authors have overall successfully created the ideal balance between theoretical presentation and practical challenges.
This book gives you a step-by-step introduction to analysing time series using the open source software R. Each time series model is motivated with practical applications, and is defined in mathematical notation. Once the model has been introduced it is used to generate synthetic data, using R code, and these generated data are then used to estimate its parameters. This sequence enhances understanding of both the time series model and the R function used to fit the model to data. Finally, the model is used to analyse observed data taken from a practical application. By using R, the whole procedure can be reproduced by the reader. All the data sets used in the book are available on the website http://staff.elena.aut.ac.nz/Paul-Cowpertwait/ts/. The book is written for undergraduate students of mathematics, economics, business and finance, geography, engineering and related disciplines, and postgraduate students who may need to analyse time series as part of their taught programme or their research.
Some of the key mathematical results are stated without proof in order to make the underlying theory acccessible to a wider audience. The book assumes a knowledge only of basic calculus, matrix algebra, and elementary statistics. The emphasis is on methods and the analysis of data sets. The logic and tools of model-building for stationary and non-stationary time series are developed in detail and numerous exercises, many of which make use of the included computer package, provide the reader with ample opportunity to develop skills in this area. The core of the book covers stationary processes, ARMA and ARIMA processes, multivariate time series and state-space models, with an optional chapter on spectral analysis. Additional topics include harmonic regression, the Burg and Hannan-Rissanen algorithms, unit roots, regression with ARMA errors, structural models, the EM algorithm, generalized state-space models with applications to time series of count data, exponential smoothing, the Holt-Winters and ARAR forecasting algorithms, transfer function models and intervention analysis. Brief introducitons are also given to cointegration and to non-linear, continuous-time and long-memory models. The time series package included in the back of the book is a slightly modified version of the package ITSM, published separately as ITSM for Windows, by Springer-Verlag, 1994. It does not handle such large data sets as ITSM for Windows, but like the latter, runs on IBM-PC compatible computers under either DOS or Windows (version 3.1 or later). The programs are all menu-driven so that the reader can immediately apply the techniques in the book to time series data, with a minimal investment of time in the computational and algorithmic aspects of the analysis.