Papers presented at the International Workshop and Conference on Combinatorial Matrix Theory & Generalized Inverses of Matrices, organized by Department of Statistics, Manipal University during 2-7 January 2012.
The sparse backslash book. Everything you wanted to know but never dared to ask about modern direct linear solvers. Chen Greif, Assistant Professor, Department of Computer Science, University of British Columbia.Overall, the book is magnificent. It fills a long-felt need for an accessible textbook on modern sparse direct methods. Its choice of scope is excellent John Gilbert, Professor, Department of Computer Science, University of California, Santa Barbara.Computational scientists often encounter problems requiring the solution of sparse systems of linear equations. Attacking these problems efficiently requires an in-depth knowledge of the underlying theory, algorithms, and data structures found in sparse matrix software libraries. Here, Davis presents the fundamentals of sparse matrix algorithms to provide the requisite background. The book includes CSparse, a concise downloadable sparse matrix package that illustrates the algorithms and theorems presented in the book and equips readers with the tools necessary to understand larger and more complex software packages.With a strong emphasis on MATLAB and the C programming language, Direct Methods for Sparse Linear Systems equips readers with the working knowledge required to use sparse solver packages and write code to interface applications to those packages. The book also explains how MATLAB performs its sparse matrix computations.Audience This invaluable book is essential to computational scientists and software developers who want to understand the theory and algorithms behind modern techniques used to solve large sparse linear systems. The book also serves as an excellent practical resource for students with an interest in combinatorial scientific computing.Preface; Chapter 1: Introduction; Chapter 2: Basic algorithms; Chapter 3: Solving triangular systems; Chapter 4: Cholesky factorization; Chapter 5: Orthogonal methods; Chapter 6: LU factorization; Chapter 7: Fill-reducing orderings; Chapter 8: Solving sparse linear systems; Chapter 9: CSparse; Chapter 10: Sparse matrices in MATLAB; Appendix: Basics of the C programming language; Bibliography; Index.
Graphs and matrices enjoy a fascinating and mutually beneficial relationship. This interplay has benefited both graph theory and linear algebra. In one direction, knowledge about one of the graphs that can be associated with a matrix can be used to illuminate matrix properties and to get better information about the matrix. Examples include the use of digraphs to obtain strong results on diagonal dominance and eigenvalue inclusion regions and the use of the Rado-Hall theorem to deduce properties of special classes of matrices. Going the other way, linear algebraic properties of one of the matrices associated with a graph can be used to obtain useful combinatorial information about the graph. The adjacency matrix and the Laplacian matrix are two well-known matrices associated to a graph, and their eigenvalues encode important information about the graph. Another important linear algebraic invariant associated with a graph is the Colin de Verdiere number, which, for instance, characterizes certain topological properties of the graph. This book is not a comprehensive study of graphs and matrices. The particular content of the lectures was chosen for its accessibility, beauty, and current relevance, and for the possibility of enticing the audience to want to learn more.
Graph-structured data is ubiquitous throughout the natural and social sciences, from telecommunication networks to quantum chemistry. Building relational inductive biases into deep learning architectures is crucial for creating systems that can learn, reason, and generalize from this kind of data. Recent years have seen a surge in research on graph representation learning, including techniques for deep graph embeddings, generalizations of convolutional neural networks to graph-structured data, and neural message-passing approaches inspired by belief propagation. These advances in graph representation learning have led to new state-of-the-art results in numerous domains, including chemical synthesis, 3D vision, recommender systems, question answering, and social network analysis. This book provides a synthesis and overview of graph representation learning. It begins with a discussion of the goals of graph representation learning as well as key methodological foundations in graph theory and network analysis. Following this, the book introduces and reviews methods for learning node embeddings, including random-walk-based methods and applications to knowledge graphs. It then provides a technical synthesis and introduction to the highly successful graph neural network (GNN) formalism, which has become a dominant and fast-growing paradigm for deep learning with graph data. The book concludes with a synthesis of recent advancements in deep generative models for graphs—a nascent but quickly growing subset of graph representation learning.
Combinatorics and Matrix Theory have a symbiotic, or mutually beneficial, relationship. This relationship is discussed in my paper The symbiotic relationship of combinatorics and matrix theoryl where I attempted to justify this description. One could say that a more detailed justification was given in my book with H. J. Ryser entitled Combinatorial Matrix Theon? where an attempt was made to give a broad picture of the use of combinatorial ideas in matrix theory and the use of matrix theory in proving theorems which, at least on the surface, are combinatorial in nature. In the book by Liu and Lai, this picture is enlarged and expanded to include recent developments and contributions of Chinese mathematicians, many of which have not been readily available to those of us who are unfamiliar with Chinese journals. Necessarily, there is some overlap with the book Combinatorial Matrix Theory. Some of the additional topics include: spectra of graphs, eulerian graph problems, Shannon capacity, generalized inverses of Boolean matrices, matrix rearrangements, and matrix completions. A topic to which many Chinese mathematicians have made substantial contributions is the combinatorial analysis of powers of nonnegative matrices, and a large chapter is devoted to this topic. This book should be a valuable resource for mathematicians working in the area of combinatorial matrix theory. Richard A. Brualdi University of Wisconsin - Madison 1 Linear Alg. Applies., vols. 162-4, 1992, 65-105 2Camhridge University Press, 1991.
Examines partitions and covers of graphs and digraphs, latin squares, pairwise balanced designs with prescribed block sizes, ranks and permanents, extremal graph theory, Hadamard matrices and graph factorizations. This book is designed to be of interest to applied mathematicians, computer scientists and communications researchers.
This new edition illustrates the power of linear algebra in the study of graphs. The emphasis on matrix techniques is greater than in other texts on algebraic graph theory. Important matrices associated with graphs (for example, incidence, adjacency and Laplacian matrices) are treated in detail. Presenting a useful overview of selected topics in algebraic graph theory, early chapters of the text focus on regular graphs, algebraic connectivity, the distance matrix of a tree, and its generalized version for arbitrary graphs, known as the resistance matrix. Coverage of later topics include Laplacian eigenvalues of threshold graphs, the positive definite completion problem and matrix games based on a graph. Such an extensive coverage of the subject area provides a welcome prompt for further exploration. The inclusion of exercises enables practical learning throughout the book. In the new edition, a new chapter is added on the line graph of a tree, while some results in Chapter 6 on Perron-Frobenius theory are reorganized. Whilst this book will be invaluable to students and researchers in graph theory and combinatorial matrix theory, it will also benefit readers in the sciences and engineering.
Graphs naturally represent information ranging from links between web pages, to communication in email networks, to connections between neurons in our brains. These graphs often span billions of nodes and interactions between them. Within this deluge of interconnected data, how can we find the most important structures and summarize them? How can we efficiently visualize them? How can we detect anomalies that indicate critical events, such as an attack on a computer system, disease formation in the human brain, or the fall of a company? This book presents scalable, principled discovery algorithms that combine globality with locality to make sense of one or more graphs. In addition to fast algorithmic methodologies, we also contribute graph-theoretical ideas and models, and real-world applications in two main areas: Individual Graph Mining: We show how to interpretably summarize a single graph by identifying its important graph structures. We complement summarization with inference, which leverages information about few entities (obtained via summarization or other methods) and the network structure to efficiently and effectively learn information about the unknown entities. Collective Graph Mining: We extend the idea of individual-graph summarization to time-evolving graphs, and show how to scalably discover temporal patterns. Apart from summarization, we claim that graph similarity is often the underlying problem in a host of applications where multiple graphs occur (e.g., temporal anomaly detection, discovery of behavioral patterns), and we present principled, scalable algorithms for aligning networks and measuring their similarity. The methods that we present in this book leverage techniques from diverse areas, such as matrix algebra, graph theory, optimization, information theory, machine learning, finance, and social science, to solve real-world problems. We present applications of our exploration algorithms to massive datasets, including a Web graph of 6.6 billion edges, a Twitter graph of 1.8 billion edges, brain graphs with up to 90 million edges, collaboration, peer-to-peer networks, browser logs, all spanning millions of users and interactions.