At the beginning of the new millennium, two unstoppable processes aretaking place in the world: (1) globalization of the economy; (2)information revolution. As a consequence, there is greaterparticipation of the world population in capital market investment, such as bonds and stocks and their derivatives
At the beginning of the new millennium, two unstoppable processes are taking place in the world: (1) globalization of the economy; (2) information revolution. As a consequence, there is greater participation of the world population in capital market investment, such as bonds and stocks and their derivatives. Hence there is a need for risk management and analytic theory explaining the market. This leads to quantitative tools based on mathematical methods, i.e. the theory of mathematical finance.Ever since the pioneer work of Black, Scholes and Merton in the 70's, there has been rapid growth in the study of mathematical finance, involving ever more sophisticated mathematics. However, from the practitioner's point of view, it is desirable to have simpler and more useful mathematical tools.This book introduces research students and practitioners to the intuitive but rigorous hypermodel techniques in finance. It is based on Robinson's infinitesimal analysis, which is easily grasped by anyone with as little background as first-year calculus. It covers topics such as pricing derivative securities (including the Black-Scholes formula), hedging, term structure models of interest rates, consumption and equilibrium. The reader is introduced to mathematical tools needed for the aforementioned topics. Mathematical proofs and details are given in an appendix. Some programs in MATHEMATICA are also included.
This book contains several innovative models for the prices of financial assets. First published in 1986, it is a classic text in the area of financial econometrics. It presents ARCH and stochastic volatility models that are often used and cited in academic research and are applied by quantitative analysts in many banks. Another often-cited contribution of the first edition is the documentation of statistical characteristics of financial returns, which are referred to as stylized facts. This second edition takes into account the remarkable progress made by empirical researchers during the past two decades from 1986 to 2006. In the new Preface, the author summarizes this progress in two key areas: firstly, measuring, modelling and forecasting volatility; and secondly, detecting and exploiting price trends. Sample Chapter(s). Chapter 1: Introduction (1,134 KB). Contents: Features of Financial Returns; Modelling Price Volatility; Forecasting Standard Deviations; The Accuracy of Autocorrelation Estimates; Testing the Random Walk Hypothesis; Forecasting Trends in Prices; Evidence Against the Efficiency of Futures Markets; Valuing Options; Appendix: A Computer Program for Modelling Financial Time Series. Readership: Academic researchers in finance & economics; quantitative analysts.
During 1995 the Isaac Newton Institute for the Mathematical Sciences at Cambridge University hosted a six month research program on financial mathematics. During this period more than 300 scholars and financial practitioners attended to conduct research and to attend more than 150 research seminars. Many of the presented papers were on the subject of financial derivatives. The very best were selected to appear in this volume. They range from abstract financial theory to practical issues pertaining to the pricing and hedging of interest rate derivatives and exotic options in the market place. Hence this book will be of interest to both academic scholars and financial engineers.
Translated from the French, this book is an introduction to first-order model theory. Starting from scratch, it quickly reaches the essentials, namely, the back-and-forth method and compactness, which are illustrated with examples taken from algebra. It also introduces logic via the study of the models of arithmetic, and it gives complete but accessible exposition of stability theory.
The book presents an axiomatic approach to the problems of prediction, classification, and statistical learning. Using methodologies from axiomatic decision theory, and, in particular, the authors' case-based decision theory, the present studies attempt to ask what inductive conclusions can be derived from existing databases. It is shown that simple consistency rules lead to similarity-weighted aggregation, akin to kernel-based methods. It is suggested that the similarity function be estimated from the data. The incorporation of rule-based reasoning is discussed.
This is an introduction to set theory and logic that starts completely from scratch. The text is accompanied by many methodological remarks and explanations. A rigorous axiomatic presentation of Zermelo-Fraenkel set theory is given, demonstrating how the basic concepts of mathematics have apparently been reduced to set theory. This is followed by a presentation of propositional and first-order logic. Concepts and results of recursion theory are explained in intuitive terms, and the author proves and explains the limitative results of Skolem, Tarski, Church and Gödel (the celebrated incompleteness theorems). For students of mathematics or philosophy this book provides an excellent introduction to logic and set theory.
In the early 1960s, by using techniques from the model theory of first-order logic, Robinson gave a rigorous formulation and extension of Leibniz'' infinitesimal calculus. Since then, the methodology has found applications in a wide spectrum of areas in mathematics, with particular success in the probability theory and functional analysis. In the latter, fruitful results were produced with Luxemburg''s invention of the nonstandard hull construction. However, there is still no publication of a coherent and self-contained treatment of functional analysis using methods from nonstandard analysis. This publication aims to fill this gap.