Machine learning and data mining are inseparably connected with uncertainty. The observable data for learning is usually imprecise, incomplete or noisy. Uncertainty Modeling for Data Mining: A Label Semantics Approach introduces 'label semantics', a fuzzy-logic-based theory for modeling uncertainty. Several new data mining algorithms based on label semantics are proposed and tested on real-world datasets. A prototype interpretation of label semantics and new prototype-based data mining algorithms are also discussed. This book offers a valuable resource for postgraduates, researchers and other professionals in the fields of data mining, fuzzy computing and uncertainty reasoning. Zengchang Qin is an associate professor at the School of Automation Science and Electrical Engineering, Beihang University, China; Yongchuan Tang is an associate professor at the College of Computer Science, Zhejiang University, China.
Modeling Uncertainty in the Earth Sciences highlights the various issues, techniques and practical modeling tools available for modeling the uncertainty of complex Earth systems and the impact that it has on practical situations. The aim of the book is to provide an introductory overview which covers a broad range of tried-and-tested tools. Descriptions of concepts, philosophies, challenges, methodologies and workflows give the reader an understanding of the best way to make decisions under uncertainty for Earth Science problems. The book covers key issues such as: Spatial and time aspect; large complexity and dimensionality; computation power; costs of 'engineering' the Earth; uncertainty in the modeling and decision process. Focusing on reliable and practical methods this book provides an invaluable primer for the complex area of decision making with uncertainty in the Earth Sciences.
This book commemorates the 65th birthday of Dr. Boris Kovalerchuk, and reflects many of the research areas covered by his work. It focuses on data processing under uncertainty, especially fuzzy data processing, when uncertainty comes from the imprecision of expert opinions. The book includes 17 authoritative contributions by leading experts.
The amount of new information is constantly increasing, faster than our ability to fully interpret and utilize it to improve human experiences. Addressing this asymmetry requires novel and revolutionary scientific methods and effective human and artificial intelligence interfaces. By lifting the concept of time from a positive real number to a 2D complex time (kime), this book uncovers a connection between artificial intelligence (AI), data science, and quantum mechanics. It proposes a new mathematical foundation for data science based on raising the 4D spacetime to a higher dimension where longitudinal data (e.g., time-series) are represented as manifolds (e.g., kime-surfaces). This new framework enables the development of innovative data science analytical methods for model-based and model-free scientific inference, derived computed phenotyping, and statistical forecasting. The book provides a transdisciplinary bridge and a pragmatic mechanism to translate quantum mechanical principles, such as particles and wavefunctions, into data science concepts, such as datum and inference-functions. It includes many open mathematical problems that still need to be solved, technological challenges that need to be tackled, and computational statistics algorithms that have to be fully developed and validated. Spacekime analytics provide mechanisms to effectively handle, process, and interpret large, heterogeneous, and continuously-tracked digital information from multiple sources. The authors propose computational methods, probability model-based techniques, and analytical strategies to estimate, approximate, or simulate the complex time phases (kime directions). This allows transforming time-varying data, such as time-series observations, into higher-dimensional manifolds representing complex-valued and kime-indexed surfaces (kime-surfaces). The book includes many illustrations of model-based and model-free spacekime analytic techniques applied to economic forecasting, identification of functional brain activation, and high-dimensional cohort phenotyping. Specific case-study examples include unsupervised clustering using the Michigan Consumer Sentiment Index (MCSI), model-based inference using functional magnetic resonance imaging (fMRI) data, and model-free inference using the UK Biobank data archive. The material includes mathematical, inferential, computational, and philosophical topics such as Heisenberg uncertainty principle and alternative approaches to large sample theory, where a few spacetime observations can be amplified by a series of derived, estimated, or simulated kime-phases. The authors extend Newton-Leibniz calculus of integration and differentiation to the spacekime manifold and discuss possible solutions to some of the "problems of time". The coverage also includes 5D spacekime formulations of classical 4D spacetime mathematical equations describing natural laws of physics, as well as, statistical articulation of spacekime analytics in a Bayesian inference framework. The steady increase of the volume and complexity of observed and recorded digital information drives the urgent need to develop novel data analytical strategies. Spacekime analytics represents one new data-analytic approach, which provides a mechanism to understand compound phenomena that are observed as multiplex longitudinal processes and computationally tracked by proxy measures. This book may be of interest to academic scholars, graduate students, postdoctoral fellows, artificial intelligence and machine learning engineers, biostatisticians, econometricians, and data analysts. Some of the material may also resonate with philosophers, futurists, astrophysicists, space industry technicians, biomedical researchers, health practitioners, and the general public.
Uncertainties are pervasive in natural hazards, and it is crucial to develop robust and meaningful approaches to characterize and communicate uncertainties to inform modeling efforts. In this monograph we provide a broad, cross-disciplinary overview of issues relating to uncertainties faced in natural hazard and risk assessment. We introduce some basic tenets of uncertainty analysis, discuss issues related to communication and decision support, and offer numerous examples of analyses and modeling approaches that vary by context and scope. Contributors include scientists from across the full breath of the natural hazard scientific community, from those in real-time analysis of natural hazards to those in the research community from academia and government. Key themes and highlights include: Substantial breadth and depth of analysis in terms of the types of natural hazards addressed, the disciplinary perspectives represented, and the number of studies included Targeted, application-centered analyses with a focus on development and use of modeling techniques to address various sources of uncertainty Emphasis on the impacts of climate change on natural hazard processes and outcomes Recommendations for cross-disciplinary and science transfer across natural hazard sciences This volume will be an excellent resource for those interested in the current work on uncertainty classification/quantification and will document common and emergent research themes to allow all to learn from each other and build a more connected but still diverse and ever growing community of scientists. Read an interview with the editors to find out more: https://eos.org/editors-vox/reducing-uncertainty-in-hazard-prediction
Probability is the bedrock of machine learning. You cannot develop a deep understanding and application of machine learning without it. Cut through the equations, Greek letters, and confusion, and discover the topics in probability that you need to know. Using clear explanations, standard Python libraries, and step-by-step tutorial lessons, you will discover the importance of probability to machine learning, Bayesian probability, entropy, density estimation, maximum likelihood, and much more.
The first truly interdisciplinary text on data mining, blending the contributions of information science, computer science, and statistics. The growing interest in data mining is motivated by a common problem across disciplines: how does one store, access, model, and ultimately describe and understand very large data sets? Historically, different aspects of data mining have been addressed independently by different disciplines. This is the first truly interdisciplinary text on data mining, blending the contributions of information science, computer science, and statistics. The book consists of three sections. The first, foundations, provides a tutorial overview of the principles underlying data mining algorithms and their application. The presentation emphasizes intuition rather than rigor. The second section, data mining algorithms, shows how algorithms are constructed to solve specific problems in a principled manner. The algorithms covered include trees and rules for classification and regression, association rules, belief networks, classical statistical models, nonlinear models such as neural networks, and local "memory-based" models. The third section shows how all of the preceding analysis fits together when applied to real-world data mining problems. Topics include the role of metadata, how to handle missing data, and data preprocessing.
This book develops a framework that shows how uncertainty in Artificial Intelligence (AI) expands and generalizes traditional AI. It explores the uncertainties of knowledge and intelligence. The authors focus on the importance of natural language – the carrier of knowledge and intelligence, and introduce efficient physical methods for data mining amd control. In this new edition, we have more in-depth description of the models and methods, of which the mathematical properties are proved strictly which make these theories and methods more complete. The authors also highlight their latest research results.
To honour the remarkable contribution of Michel David in the inception, establishment and development of Geostatistics, and to promote the essence of his work, an international Forum entitled Geostatistics for the Next Century was convened in Montreal in June 1993. In order to enhance communication and stimulate geostatistical innovation, research and development, the Forum brought together world leading researchers and practitioners from five continents, who discussed-debated current problems, new technologies and futuristic ideas. This volume contains selected peer-reviewed papers from the Forum, together with comments by participants and replies by authors. Although difficult to capture the spontaneity and range of a debate, comments and replies should further assist in the promotion of ideas, dialogue and criticism, and are consistent with the spirit of the Forum. The contents of this volume are organized following the Forum's thematic sessions. The role of theme sessions was not only to stress important topics of tOday but in addition, to emphasize common ground held among diverse areas of geostatistical work and the need to strengthen communication between these areas. For this reason, any given section of this book may include papers from theory to applications, in mining, petroleum, environment, geohydrology, image processing.
Under the Earth's surface is a rich array of geological resources, many with potential use to humankind. However, extracting and harnessing them comes with enormous uncertainties, high costs, and considerable risks. The valuation of subsurface resources involves assessing discordant factors to produce a decision model that is functional and sustainable. This volume provides real-world examples relating to oilfields, geothermal systems, contaminated sites, and aquifer recharge. Volume highlights include: A multi-disciplinary treatment of uncertainty quantification Case studies with actual data that will appeal to methodology developers A Bayesian evidential learning framework that reduces computation and modeling time Quantifying Uncertainty in Subsurface Systems is a multidisciplinary volume that brings together five major fields: information science, decision science, geosciences, data science and computer science. It will appeal to both students and practitioners, and be a valuable resource for geoscientists, engineers and applied mathematicians. Read the Editors' Vox: eos.org/editors-vox/quantifying-uncertainty-about-earths-resources