Comprehensive introduction to the neural network models currently under intensive study for computational applications. It also provides coverage of neural network applications in a variety of problems of both theoretical and practical interest.
The new edition of Fundamentals of Computational Neuroscience build on the success and strengths of the first edition. Completely redesigned and revised, it introduces the theoretical foundations of neuroscience with a focus on the nature of information processing in the brain.
A textbook for students with limited background in mathematics and computer coding, emphasizing computer tutorials that guide readers in producing models of neural behavior. This introductory text teaches students to understand, simulate, and analyze the complex behaviors of individual neurons and brain circuits. It is built around computer tutorials that guide students in producing models of neural behavior, with the associated Matlab code freely available online. From these models students learn how individual neurons function and how, when connected, neurons cooperate in a circuit. The book demonstrates through simulated models how oscillations, multistability, post-stimulus rebounds, and chaos can arise within either single neurons or circuits, and it explores their roles in the brain. The book first presents essential background in neuroscience, physics, mathematics, and Matlab, with explanations illustrated by many example problems. Subsequent chapters cover the neuron and spike production; single spike trains and the underlying cognitive processes; conductance-based models; the simulation of synaptic connections; firing-rate models of large-scale circuit operation; dynamical systems and their components; synaptic plasticity; and techniques for analysis of neuron population datasets, including principal components analysis, hidden Markov modeling, and Bayesian decoding. Accessible to undergraduates in life sciences with limited background in mathematics and computer coding, the book can be used in a “flipped” or “inverted” teaching approach, with class time devoted to hands-on work on the computer tutorials. It can also be a resource for graduate students in the life sciences who wish to gain computing skills and a deeper knowledge of neural function and neural circuits.
This book brings together a selection of papers by George Gerstein, representing his long-term endeavor of making neuroscience into a more rigorous science inspired by physics, where he had his roots. Professor Gerstein was many years ahead of the field, consistently striving for quantitative analyses, mechanistic models, and conceptual clarity. In doing so, he pioneered Computational Neuroscience, many years before the term itself was born. The overarching goal of George Gerstein’s research was to understand the functional organization of neuronal networks in the brain. The editors of this book have compiled a selection of George Gerstein’s many seminal contributions to neuroscience--be they experimental, theoretical or computational--into a single, comprehensive volume .The aim is to provide readers with a fresh introduction of these various concepts in the original literature. The volume is organized in a series of chapters by subject, ordered in time, each one containing one or more of George Gerstein’s papers.
"For the neuroscientist or psychologist who cringes at the sight of mathematical formulae and whose eyes glaze over at terms like differential equations, linear algebra, vectors, matrices, Bayes’ rule, and Boolean logic, this book just might be the therapy needed." - Anjan Chatterjee, Professor of Neurology, University of Pennsylvania "Anderson provides a gentle introduction to computational aspects of psychological science, managing to respect the reader’s intelligence while also being completely unintimidating. Using carefully-selected computational demonstrations, he guides students through a wide array of important approaches and tools, with little in the way of prerequisites...I recommend it with enthusiasm." - Asohan Amarasingham, The City University of New York This unique, self-contained and accessible textbook provides an introduction to computational modelling neuroscience accessible to readers with little or no background in computing or mathematics. Organized into thematic sections, the book spans from modelling integrate and firing neurons to playing the game Rock, Paper, Scissors in ACT-R. This non-technical guide shows how basic knowledge and modern computers can be combined for interesting simulations, progressing from early exercises utilizing spreadsheets, to simple programs in Python. Key Features include: Interleaved chapters that show how traditional computing constructs are simply disguised versions of the spread sheet methods. Mathematical facts and notation needed to understand the modelling methods are presented at their most basic and are interleaved with biographical and historical notes for contex. Numerous worked examples to demonstrate the themes and procedures of cognitive modelling. An excellent text for postgraduate students taking courses in research methods, computational neuroscience, computational modelling, cognitive science and neuroscience. It will be especially valuable to psychology students.
Computational Neuroscience - A First Course provides an essential introduction to computational neuroscience and equips readers with a fundamental understanding of modeling the nervous system at the membrane, cellular, and network level. The book, which grew out of a lecture series held regularly for more than ten years to graduate students in neuroscience with backgrounds in biology, psychology and medicine, takes its readers on a journey through three fundamental domains of computational neuroscience: membrane biophysics, systems theory and artificial neural networks. The required mathematical concepts are kept as intuitive and simple as possible throughout the book, making it fully accessible to readers who are less familiar with mathematics. Overall, Computational Neuroscience - A First Course represents an essential reference guide for all neuroscientists who use computational methods in their daily work, as well as for any theoretical scientist approaching the field of computational neuroscience.
This text, based on a course taught by Randall O'Reilly and Yuko Munakata over the past several years, provides an in-depth introduction to the main ideas in the computational cognitive neuroscience. The goal of computational cognitive neuroscience is to understand how the brain embodies the mind by using biologically based computational models comprising networks of neuronlike units. This text, based on a course taught by Randall O'Reilly and Yuko Munakata over the past several years, provides an in-depth introduction to the main ideas in the field. The neural units in the simulations use equations based directly on the ion channels that govern the behavior of real neurons, and the neural networks incorporate anatomical and physiological properties of the neocortex. Thus the text provides the student with knowledge of the basic biology of the brain as well as the computational skills needed to simulate large-scale cognitive phenomena. The text consists of two parts. The first part covers basic neural computation mechanisms: individual neurons, neural networks, and learning mechanisms. The second part covers large-scale brain area organization and cognitive phenomena: perception and attention, memory, language, and higher-level cognition. The second part is relatively self-contained and can be used separately for mechanistically oriented cognitive neuroscience courses. Integrated throughout the text are more than forty different simulation models, many of them full-scale research-grade models, with friendly interfaces and accompanying exercises. The simulation software (PDP++, available for all major platforms) and simulations can be downloaded free of charge from the Web. Exercise solutions are available, and the text includes full information on the software.
This book provides a comprehensive introduction to the computational material that forms the underpinnings of the currently evolving set of brain models. It is now clear that the brain is unlikely to be understood without recourse to computational theories. The theme of An Introduction to Natural Computation is that ideas from diverse areas such as neuroscience, information theory, and optimization theory have recently been extended in ways that make them useful for describing the brains programs. This book provides a comprehensive introduction to the computational material that forms the underpinnings of the currently evolving set of brain models. It stresses the broad spectrum of learning models—ranging from neural network learning through reinforcement learning to genetic learning—and situates the various models in their appropriate neural context. To write about models of the brain before the brain is fully understood is a delicate matter. Very detailed models of the neural circuitry risk losing track of the task the brain is trying to solve. At the other extreme, models that represent cognitive constructs can be so abstract that they lose all relationship to neurobiology. An Introduction to Natural Computation takes the middle ground and stresses the computational task while staying near the neurobiology.
Neural network research often builds on the fiction that neurons are simple linear threshold units, completely neglecting the highly dynamic and complex nature of synapses, dendrites, and voltage-dependent ionic currents. Biophysics of Computation: Information Processing in Single Neurons challenges this notion, using richly detailed experimental and theoretical findings from cellular biophysics to explain the repertoire of computational functions available to single neurons. The author shows how individual nerve cells can multiply, integrate, or delay synaptic inputs and how information can be encoded in the voltage across the membrane, in the intracellular calcium concentration, or in the timing of individual spikes.Key topics covered include the linear cable equation; cable theory as applied to passive dendritic trees and dendritic spines; chemical and electrical synapses and how to treat them from a computational point of view; nonlinear interactions of synaptic input in passive and active dendritic trees; the Hodgkin-Huxley model of action potential generation and propagation; phase space analysis; linking stochastic ionic channels to membrane-dependent currents; calcium and potassium currents and their role in information processing; the role of diffusion, buffering and binding of calcium, and other messenger systems in information processing and storage; short- and long-term models of synaptic plasticity; simplified models of single cells; stochastic aspects of neuronal firing; the nature of the neuronal code; and unconventional models of sub-cellular computation.Biophysics of Computation: Information Processing in Single Neurons serves as an ideal text for advanced undergraduate and graduate courses in cellular biophysics, computational neuroscience, and neural networks, and will appeal to students and professionals in neuroscience, electrical and computer engineering, and physics.