Space-Time Computing with Temporal Neural Networks

Space-Time Computing with Temporal Neural Networks

Author: James E. Smith

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 220

ISBN-13: 3031017544

DOWNLOAD EBOOK

Understanding and implementing the brain's computational paradigm is the one true grand challenge facing computer researchers. Not only are the brain's computational capabilities far beyond those of conventional computers, its energy efficiency is truly remarkable. This book, written from the perspective of a computer designer and targeted at computer researchers, is intended to give both background and lay out a course of action for studying the brain's computational paradigm. It contains a mix of concepts and ideas drawn from computational neuroscience, combined with those of the author. As background, relevant biological features are described in terms of their computational and communication properties. The brain's neocortex is constructed of massively interconnected neurons that compute and communicate via voltage spikes, and a strong argument can be made that precise spike timing is an essential element of the paradigm. Drawing from the biological features, a mathematics-based computational paradigm is constructed. The key feature is spiking neurons that perform communication and processing in space-time, with emphasis on time. In these paradigms, time is used as a freely available resource for both communication and computation. Neuron models are first discussed in general, and one is chosen for detailed development. Using the model, single-neuron computation is first explored. Neuron inputs are encoded as spike patterns, and the neuron is trained to identify input pattern similarities. Individual neurons are building blocks for constructing larger ensembles, referred to as "columns". These columns are trained in an unsupervised manner and operate collectively to perform the basic cognitive function of pattern clustering. Similar input patterns are mapped to a much smaller set of similar output patterns, thereby dividing the input patterns into identifiable clusters. Larger cognitive systems are formed by combining columns into a hierarchical architecture. These higher level architectures are the subject of ongoing study, and progress to date is described in detail in later chapters. Simulation plays a major role in model development, and the simulation infrastructure developed by the author is described.


Space-Time Computing with Temporal Neural Networks

Space-Time Computing with Temporal Neural Networks

Author: James E. Smith

Publisher: Morgan & Claypool Publishers

Published: 2017-05-18

Total Pages: 245

ISBN-13: 1627058907

DOWNLOAD EBOOK

Understanding and implementing the brain's computational paradigm is the one true grand challenge facing computer researchers. Not only are the brain's computational capabilities far beyond those of conventional computers, its energy efficiency is truly remarkable. This book, written from the perspective of a computer designer and targeted at computer researchers, is intended to give both background and lay out a course of action for studying the brain's computational paradigm. It contains a mix of concepts and ideas drawn from computational neuroscience, combined with those of the author. As background, relevant biological features are described in terms of their computational and communication properties. The brain's neocortex is constructed of massively interconnected neurons that compute and communicate via voltage spikes, and a strong argument can be made that precise spike timing is an essential element of the paradigm. Drawing from the biological features, a mathematics-based computational paradigm is constructed. The key feature is spiking neurons that perform communication and processing in space-time, with emphasis on time. In these paradigms, time is used as a freely available resource for both communication and computation. Neuron models are first discussed in general, and one is chosen for detailed development. Using the model, single-neuron computation is first explored. Neuron inputs are encoded as spike patterns, and the neuron is trained to identify input pattern similarities. Individual neurons are building blocks for constructing larger ensembles, referred to as "columns". These columns are trained in an unsupervised manner and operate collectively to perform the basic cognitive function of pattern clustering. Similar input patterns are mapped to a much smaller set of similar output patterns, thereby dividing the input patterns into identifiable clusters. Larger cognitive systems are formed by combining columns into a hierarchical architecture. These higher level architectures are the subject of ongoing study, and progress to date is described in detail in later chapters. Simulation plays a major role in model development, and the simulation infrastructure developed by the author is described.


Time-Space, Spiking Neural Networks and Brain-Inspired Artificial Intelligence

Time-Space, Spiking Neural Networks and Brain-Inspired Artificial Intelligence

Author: Nikola K. Kasabov

Publisher: Springer

Published: 2018-08-29

Total Pages: 742

ISBN-13: 3662577151

DOWNLOAD EBOOK

Spiking neural networks (SNN) are biologically inspired computational models that represent and process information internally as trains of spikes. This monograph book presents the classical theory and applications of SNN, including original author’s contribution to the area. The book introduces for the first time not only deep learning and deep knowledge representation in the human brain and in brain-inspired SNN, but takes that further to develop new types of AI systems, called in the book brain-inspired AI (BI-AI). BI-AI systems are illustrated on: cognitive brain data, including EEG, fMRI and DTI; audio-visual data; brain-computer interfaces; personalized modelling in bio-neuroinformatics; multisensory streaming data modelling in finance, environment and ecology; data compression; neuromorphic hardware implementation. Future directions, such as the integration of multiple modalities, such as quantum-, molecular- and brain information processing, is presented in the last chapter. The book is a research book for postgraduate students, researchers and practitioners across wider areas, including computer and information sciences, engineering, applied mathematics, bio- and neurosciences.


Efficient Processing of Deep Neural Networks

Efficient Processing of Deep Neural Networks

Author: Vivienne Sze

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 254

ISBN-13: 3031017668

DOWNLOAD EBOOK

This book provides a structured treatment of the key principles and techniques for enabling efficient processing of deep neural networks (DNNs). DNNs are currently widely used for many artificial intelligence (AI) applications, including computer vision, speech recognition, and robotics. While DNNs deliver state-of-the-art accuracy on many AI tasks, it comes at the cost of high computational complexity. Therefore, techniques that enable efficient processing of deep neural networks to improve key metrics—such as energy-efficiency, throughput, and latency—without sacrificing accuracy or increasing hardware costs are critical to enabling the wide deployment of DNNs in AI systems. The book includes background on DNN processing; a description and taxonomy of hardware architectural approaches for designing DNN accelerators; key metrics for evaluating and comparing different designs; features of DNN processing that are amenable to hardware/algorithm co-design to improve energy efficiency and throughput; and opportunities for applying new technologies. Readers will find a structured introduction to the field as well as formalization and organization of key concepts from contemporary work that provide insights that may spark new ideas.


Artificial Intelligence in the Age of Neural Networks and Brain Computing

Artificial Intelligence in the Age of Neural Networks and Brain Computing

Author: Robert Kozma

Publisher: Academic Press

Published: 2023-10-11

Total Pages: 398

ISBN-13: 0323958168

DOWNLOAD EBOOK

Artificial Intelligence in the Age of Neural Networks and Brain Computing, Second Edition demonstrates that present disruptive implications and applications of AI is a development of the unique attributes of neural networks, mainly machine learning, distributed architectures, massive parallel processing, black-box inference, intrinsic nonlinearity, and smart autonomous search engines. The book covers the major basic ideas of "brain-like computing" behind AI, provides a framework to deep learning, and launches novel and intriguing paradigms as possible future alternatives. The present success of AI-based commercial products proposed by top industry leaders, such as Google, IBM, Microsoft, Intel, and Amazon, can be interpreted using the perspective presented in this book by viewing the co-existence of a successful synergism among what is referred to as computational intelligence, natural intelligence, brain computing, and neural engineering. The new edition has been updated to include major new advances in the field, including many new chapters. - Developed from the 30th anniversary of the International Neural Network Society (INNS) and the 2017 International Joint Conference on Neural Networks (IJCNN - Authored by top experts, global field pioneers, and researchers working on cutting-edge applications in signal processing, speech recognition, games, adaptive control and decision-making - Edited by high-level academics and researchers in intelligent systems and neural networks - Includes all new chapters, including topics such as Frontiers in Recurrent Neural Network Research; Big Science, Team Science, Open Science for Neuroscience; A Model-Based Approach for Bridging Scales of Cortical Activity; A Cognitive Architecture for Object Recognition in Video; How Brain Architecture Leads to Abstract Thought; Deep Learning-Based Speech Separation and Advances in AI, Neural Networks


Deep Learning for Computer Architects

Deep Learning for Computer Architects

Author: Brandon Reagen

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 109

ISBN-13: 3031017560

DOWNLOAD EBOOK

Machine learning, and specifically deep learning, has been hugely disruptive in many fields of computer science. The success of deep learning techniques in solving notoriously difficult classification and regression problems has resulted in their rapid adoption in solving real-world problems. The emergence of deep learning is widely attributed to a virtuous cycle whereby fundamental advancements in training deeper models were enabled by the availability of massive datasets and high-performance computer hardware. This text serves as a primer for computer architects in a new and rapidly evolving field. We review how machine learning has evolved since its inception in the 1960s and track the key developments leading up to the emergence of the powerful deep learning techniques that emerged in the last decade. Next we review representative workloads, including the most commonly used datasets and seminal networks across a variety of domains. In addition to discussing the workloads themselves, we also detail the most popular deep learning tools and show how aspiring practitioners can use the tools with the workloads to characterize and optimize DNNs. The remainder of the book is dedicated to the design and optimization of hardware and architectures for machine learning. As high-performance hardware was so instrumental in the success of machine learning becoming a practical solution, this chapter recounts a variety of optimizations proposed recently to further improve future designs. Finally, we present a review of recent research published in the area as well as a taxonomy to help readers understand how various contributions fall in context.


Robotic Computing on FPGAs

Robotic Computing on FPGAs

Author: Shaoshan Liu

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 202

ISBN-13: 3031017714

DOWNLOAD EBOOK

This book provides a thorough overview of the state-of-the-art field-programmable gate array (FPGA)-based robotic computing accelerator designs and summarizes their adopted optimized techniques. This book consists of ten chapters, delving into the details of how FPGAs have been utilized in robotic perception, localization, planning, and multi-robot collaboration tasks. In addition to individual robotic tasks, this book provides detailed descriptions of how FPGAs have been used in robotic products, including commercial autonomous vehicles and space exploration robots.


In-/Near-Memory Computing

In-/Near-Memory Computing

Author: Daichi Fujiki

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 124

ISBN-13: 3031017722

DOWNLOAD EBOOK

This book provides a structured introduction of the key concepts and techniques that enable in-/near-memory computing. For decades, processing-in-memory or near-memory computing has been attracting growing interest due to its potential to break the memory wall. Near-memory computing moves compute logic near the memory, and thereby reduces data movement. Recent work has also shown that certain memories can morph themselves into compute units by exploiting the physical properties of the memory cells, enabling in-situ computing in the memory array. While in- and near-memory computing can circumvent overheads related to data movement, it comes at the cost of restricted flexibility of data representation and computation, design challenges of compute capable memories, and difficulty in system and software integration. Therefore, wide deployment of in-/near-memory computing cannot be accomplished without techniques that enable efficient mapping of data-intensive applications to such devices, without sacrificing accuracy or increasing hardware costs excessively. This book describes various memory substrates amenable to in- and near-memory computing, architectural approaches for designing efficient and reliable computing devices, and opportunities for in-/near-memory acceleration of different classes of applications.


AI for Computer Architecture

AI for Computer Architecture

Author: Lizhong Chen

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 124

ISBN-13: 3031017706

DOWNLOAD EBOOK

Artificial intelligence has already enabled pivotal advances in diverse fields, yet its impact on computer architecture has only just begun. In particular, recent work has explored broader application to the design, optimization, and simulation of computer architecture. Notably, machine-learning-based strategies often surpass prior state-of-the-art analytical, heuristic, and human-expert approaches. This book reviews the application of machine learning in system-wide simulation and run-time optimization, and in many individual components such as caches/memories, branch predictors, networks-on-chip, and GPUs. The book further analyzes current practice to highlight useful design strategies and identify areas for future work, based on optimized implementation strategies, opportune extensions to existing work, and ambitious long term possibilities. Taken together, these strategies and techniques present a promising future for increasingly automated computer architecture designs.


Quantum Computer Systems

Quantum Computer Systems

Author: Yongshan Ding

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 203

ISBN-13: 303101765X

DOWNLOAD EBOOK

This book targets computer scientists and engineers who are familiar with concepts in classical computer systems but are curious to learn the general architecture of quantum computing systems. It gives a concise presentation of this new paradigm of computing from a computer systems' point of view without assuming any background in quantum mechanics. As such, it is divided into two parts. The first part of the book provides a gentle overview on the fundamental principles of the quantum theory and their implications for computing. The second part is devoted to state-of-the-art research in designing practical quantum programs, building a scalable software systems stack, and controlling quantum hardware components. Most chapters end with a summary and an outlook for future directions. This book celebrates the remarkable progress that scientists across disciplines have made in the past decades and reveals what roles computer scientists and engineers can play to enable practical-scale quantum computing.