This book offers detailed surveys and systematic discussion of models, algorithms and applications for link mining, focusing on theory and technique, and related applications: text mining, social network analysis, collaborative filtering and bioinformatics.
This book offers detailed surveys and systematic discussion of models, algorithms and applications for link mining, focusing on theory and technique, and related applications: text mining, social network analysis, collaborative filtering and bioinformatics.
Markov chains are a particularly powerful and widely used tool for analyzing a variety of stochastic (probabilistic) systems over time. This monograph will present a series of Markov models, starting from the basic models and then building up to higher-order models. Included in the higher-order discussions are multivariate models, higher-order multivariate models, and higher-order hidden models. In each case, the focus is on the important kinds of applications that can be made with the class of models being considered in the current chapter. Special attention is given to numerical algorithms that can efficiently solve the models. Therefore, Markov Chains: Models, Algorithms and Applications outlines recent developments of Markov chain models for modeling queueing sequences, Internet, re-manufacturing systems, reverse logistics, inventory systems, bio-informatics, DNA sequences, genetic networks, data mining, and many other practical systems.
Network data are produced automatically by everyday interactions - social networks, power grids, and links between data sets are a few examples. Such data capture social and economic behavior in a form that can be analyzed using powerful computational tools. This book is a guide to both basic and advanced techniques and algorithms for extracting useful information from network data. The content is organized around 'tasks', grouping the algorithms needed to gather specific types of information and thus answer specific types of questions. Examples include similarity between nodes in a network, prestige or centrality of individual nodes, and dense regions or communities in a network. Algorithms are derived in detail and summarized in pseudo-code. The book is intended primarily for computer scientists, engineers, statisticians and physicists, but it is also accessible to network scientists based in the social sciences. MATLAB®/Octave code illustrating some of the algorithms will be available at: http://www.cambridge.org/9781107125773.
This volume of Advances in Intelligent and Soft Computing contains accepted papers presented at the 10th International Conference on Soft Computing Models in Industrial and Environmental Applications (SOCO 2015), held in the beautiful and historic city of Burgos (Spain), in June 2015. Soft computing represents a collection or set of computational techniques in machine learning, computer science and some engineering disciplines, which investigate, simulate and analyze very complex issues and phenomena. This Conference is mainly focused on its industrial and environmental applications. After a through peer-review process, the SOCO 2015 International Program Committee selected 41 papers, written by authors from 15 different countries. These papers are published in present conference proceedings, achieving an acceptance rate of 40%. The selection of papers was extremely rigorous in order to maintain the high quality of the conference and we would like to thank the members of the International Program Committees for their hard work during the review process. This is a crucial issue for creation of a high standard conference and the SOCO conference would not exist without their help.
Data matching (also known as record or data linkage, entity resolution, object identification, or field matching) is the task of identifying, matching and merging records that correspond to the same entities from several databases or even within one database. Based on research in various domains including applied statistics, health informatics, data mining, machine learning, artificial intelligence, database management, and digital libraries, significant advances have been achieved over the last decade in all aspects of the data matching process, especially on how to improve the accuracy of data matching, and its scalability to large databases. Peter Christen’s book is divided into three parts: Part I, “Overview”, introduces the subject by presenting several sample applications and their special challenges, as well as a general overview of a generic data matching process. Part II, “Steps of the Data Matching Process”, then details its main steps like pre-processing, indexing, field and record comparison, classification, and quality evaluation. Lastly, part III, “Further Topics”, deals with specific aspects like privacy, real-time matching, or matching unstructured data. Finally, it briefly describes the main features of many research and open source systems available today. By providing the reader with a broad range of data matching concepts and techniques and touching on all aspects of the data matching process, this book helps researchers as well as students specializing in data quality or data matching aspects to familiarize themselves with recent research advances and to identify open research challenges in the area of data matching. To this end, each chapter of the book includes a final section that provides pointers to further background and research material. Practitioners will better understand the current state of the art in data matching as well as the internal workings and limitations of current systems. Especially, they will learn that it is often not feasible to simply implement an existing off-the-shelf data matching system without substantial adaption and customization. Such practical considerations are discussed for each of the major steps in the data matching process.
This book constitutes the refereed proceedings of the 9th International Conference on Modeling Decisions for Artificial Intelligence, MDAI 2012, held in Girona, Catalonia, Spain, in November 2012. The 32 revised full papers were carefully reviewed and selected from 49 submissions and are presented with 4 plenary talks. The papers are organized in topical sections on aggregation operators, integrals, data privacy and security, reasoning, applications, and clustering and similarity.
This book constitutes the refereed proceedings of the 32nd International Conference on Conceptual Modeling, ER 2014, held in Atlanta, GA, USA. The 23 full and 15 short papers presented were carefully reviewed and selected from 80 submissions. Topics of interest presented and discussed in the conference span the entire spectrum of conceptual modeling including research and practice in areas such as: data on the web, unstructured data, uncertain and incomplete data, big data, graphs and networks, privacy and safety, database design, new modeling languages and applications, software concepts and strategies, patterns and narratives, data management for enterprise architecture, city and urban applications.