This three volume set LNCS 6352, LNCS 6353, and LNCS 6354 constitutes the refereed proceedings of the 20th International Conference on Artificial Neural Networks, ICANN 2010, held in Thessaloniki, Greece, in September 2010. The 102 revised full papers, 68 short papers and 29 posters presented were carefully reviewed and selected from 241 submissions. The first volume is divided in topical sections on ANN applications, Bayesian ANN, bio inspired – spiking ANN, biomedical ANN, computational neuroscience, feature selection/parameter identification and dimensionality reduction, filtering, genetic – evolutionary algorithms, and image – video and audio processing.
The two-volume set LNCS 7552 + 7553 constitutes the proceedings of the 22nd International Conference on Artificial Neural Networks, ICANN 2012, held in Lausanne, Switzerland, in September 2012. The 162 papers included in the proceedings were carefully reviewed and selected from 247 submissions. They are organized in topical sections named: theoretical neural computation; information and optimization; from neurons to neuromorphism; spiking dynamics; from single neurons to networks; complex firing patterns; movement and motion; from sensation to perception; object and face recognition; reinforcement learning; bayesian and echo state networks; recurrent neural networks and reservoir computing; coding architectures; interacting with the brain; swarm intelligence and decision-making; mulitlayer perceptrons and kernel networks; training and learning; inference and recognition; support vector machines; self-organizing maps and clustering; clustering, mining and exploratory analysis; bioinformatics; and time weries and forecasting.
This two volume set LNCS 6791 and LNCS 6792 constitutes the refereed proceedings of the 21th International Conference on Artificial Neural Networks, ICANN 2011, held in Espoo, Finland, in June 2011. The 106 revised full or poster papers presented were carefully reviewed and selected from numerous submissions. ICANN 2011 had two basic tracks: brain-inspired computing and machine learning research, with strong cross-disciplinary interactions and applications.
The Mind and Brain are usually considered as one and the same nonlinear, complex dynamical system, in which information processing can be described with vector and tensor transformations and with attractors in multidimensional state spaces. Thus, an internal neurocognitive representation concept consists of a dynamical process which filters out statistical prototypes from the sensorial information in terms of coherent and adaptive n-dimensional vector fields. These prototypes serve as a basis for dynamic, probabilistic predictions or probabilistic hypotheses on prospective new data (see the recently introduced approach of "predictive coding" in neurophilosophy). Furthermore, the phenomenon of sensory and language cognition would thus be based on a multitude of self-regulatory complex dynamics of synchronous self-organization mechanisms, in other words, an emergent "flux equilibrium process" ("steady state") of the total collective and coherent neural activity resulting from the oscillatory actions of neuronal assemblies. In perception it is shown how sensory object informations, like the object color or the object form, can be dynamically related together or can be integrated to a neurally based representation of this perceptual object by means of a synchronization mechanism ("feature binding"). In language processing it is shown how semantic concepts and syntactic roles can be dynamically related together or can be integrated to neurally based systematic and compositional connectionist representations by means of a synchronization mechanism ("variable binding") solving the Fodor-Pylyshyn-Challenge. Since the systemtheoretical connectionism has succeeded in modeling the sensory objects in perception as well as systematic and compositional representations in language processing with this vector- and oscillation-based representation format, a new, convincing theory of neurocognition has been developed, which bridges the neuronal and the cognitive analysis level. The book describes how elementary neuronal information is combined in perception and language, so it becomes clear how the brain processes this information to enable basic cognitive performance of the humans.
This three volume set LNCS 6352, LNCS 6353, and LNCS 6354 constitutes the refereed proceedings of the 20th International Conference on Artificial Neural Networks, ICANN 2010, held in Thessaloniki, Greece, in September 20010. The 102 revised full papers, 68 short papers and 29 posters presented were carefully reviewed and selected from 241 submissions. The third volume is divided in topical sections on classification – pattern recognition, learning algorithms and systems, computational intelligence, IEM3 workshop, CVA workshop, and SOINN workshop.
This two volume set (LNCS 6791 and LNCS 6792) constitutes the refereed proceedings of the 21th International Conference on Artificial Neural Networks, ICANN 2011, held in Espoo, Finland, in June 2011. The 106 revised full or poster papers presented were carefully reviewed and selected from numerous submissions. ICANN 2011 had two basic tracks: brain-inspired computing and machine learning research, with strong cross-disciplinary interactions and applications.
Understanding and implementing the brain's computational paradigm is the one true grand challenge facing computer researchers. Not only are the brain's computational capabilities far beyond those of conventional computers, its energy efficiency is truly remarkable. This book, written from the perspective of a computer designer and targeted at computer researchers, is intended to give both background and lay out a course of action for studying the brain's computational paradigm. It contains a mix of concepts and ideas drawn from computational neuroscience, combined with those of the author. As background, relevant biological features are described in terms of their computational and communication properties. The brain's neocortex is constructed of massively interconnected neurons that compute and communicate via voltage spikes, and a strong argument can be made that precise spike timing is an essential element of the paradigm. Drawing from the biological features, a mathematics-based computational paradigm is constructed. The key feature is spiking neurons that perform communication and processing in space-time, with emphasis on time. In these paradigms, time is used as a freely available resource for both communication and computation. Neuron models are first discussed in general, and one is chosen for detailed development. Using the model, single-neuron computation is first explored. Neuron inputs are encoded as spike patterns, and the neuron is trained to identify input pattern similarities. Individual neurons are building blocks for constructing larger ensembles, referred to as "columns". These columns are trained in an unsupervised manner and operate collectively to perform the basic cognitive function of pattern clustering. Similar input patterns are mapped to a much smaller set of similar output patterns, thereby dividing the input patterns into identifiable clusters. Larger cognitive systems are formed by combining columns into a hierarchical architecture. These higher level architectures are the subject of ongoing study, and progress to date is described in detail in later chapters. Simulation plays a major role in model development, and the simulation infrastructure developed by the author is described.
These proceedings of the SAI Intelligent Systems Conference 2016 (IntelliSys 2016) offer a remarkable collection of papers on a wide range of topics in intelligent systems, and their applications to the real world. Authors hailing from 56 countries on 5 continents submitted 404 papers to the conference, attesting to the global importance of the conference’s themes. After being reviewed, 222 papers were accepted for presentation, and 168 were ultimately selected for these proceedings. Each has been reviewed on the basis of its originality, novelty and rigorousness. The papers not only present state-of-the-art methods and valuable experience from researchers in the related research areas; they also outline the field’s future development.
This book discusses in-depth the concept of distributed artificial intelligence (DAI) and its application to cognitive communications In this book, the authors present an overview of cognitive communications, encompassing both cognitive radio and cognitive networks, and also other application areas such as cognitive acoustics. The book also explains the specific rationale for the integration of different forms of distributed artificial intelligence into cognitive communications, something which is often neglected in many forms of technical contributions available today. Furthermore, the chapters are divided into four disciplines: wireless communications, distributed artificial intelligence, regulatory policy and economics and implementation. The book contains contributions from leading experts (academia and industry) in the field. Key Features: Covers the broader field of cognitive communications as a whole, addressing application to communication systems in general (e.g. cognitive acoustics and Distributed Artificial Intelligence (DAI) Illustrates how different DAI based techniques can be used to self-organise the radio spectrum Explores the regulatory, policy and economic issues of cognitive communications in the context of secondary spectrum access Discusses application and implementation of cognitive communications techniques in different application areas (e.g. Cognitive Femtocell Networks (CFN) Written by experts in the field from both academia and industry Cognitive Communications will be an invaluable guide for research community (PhD students, researchers) in the areas of wireless communications, and development engineers involved in the design and development of mobile, portable and fixed wireless systems., wireless network design engineer. Undergraduate and postgraduate students on elective courses in electronic engineering or computer science, and the research and engineering community will also find this book of interest.