When, in 1984?86, Richard P. Feynman gave his famous course on computation at the California Institute of Technology, he asked Tony Hey to adapt his lecture notes into a book. Although led by Feynman, the course also featured, as occasional guest speakers, some of the most brilliant men in science at that time, including Marvin Minsky, Charles Bennett, and John Hopfield. Although the lectures are now thirteen years old, most of the material is timeless and presents a ?Feynmanesque? overview of many standard and some not-so-standard topics in computer science such as reversible logic gates and quantum computers.
Computational properties of use to biological organisms or to the construction of computers can emerge as collective properties of systems having a large number of simple equivalent components (or neurons). The physical meaning of content-addressable memory is described by an appropriate phase space flow of the state of a system. A model of such a system is given, based on aspects of neurobiology but readily adapted to integrated circuits. The collective properties of this model produce a content-addressable memory which correctly yields an entire memory from any subpart of sufficient size. The algorithm for the time evolution of the state of the system is based on asynchronous parallel processing. Additional emergent collective properties include some capacity for generalization, familiarity recognition, categorization, error correction, and time sequence retention. The collective properties are only weakly sensitive to details of the modeling or the failure of individual devices.
Covering the theory of computation, information and communications, the physical aspects of computation, and the physical limits of computers, this text is based on the notes taken by one of its editors, Tony Hey, on a lecture course on computation given b
This book is a collection of lecture notes from the Symposium on Quantum Computing, Thermodynamics, and Statistical Physics, held at Kinki University in March 2012. Quantum information theory has a deep connection with statistical physics and thermodynamics. This volume introduces some of the topics on interface among the mentioned fields. Subjects included in the lecture notes include quantum annealing method, nonequilibrium thermodynamics and spin glass theory, among others. These subjects were presented with much emphasis put in its relevance in quantum information theory. These lecture notes are prepared in a self-contained manner so that a reader with modest background may understand the subjects.
The new edition of an introductory text that teaches students the art of computational problem solving, covering topics ranging from simple algorithms to information visualization. This book introduces students with little or no prior programming experience to the art of computational problem solving using Python and various Python libraries, including PyLab. It provides students with skills that will enable them to make productive use of computational techniques, including some of the tools and techniques of data science for using computation to model and interpret data. The book is based on an MIT course (which became the most popular course offered through MIT's OpenCourseWare) and was developed for use not only in a conventional classroom but in in a massive open online course (MOOC). This new edition has been updated for Python 3, reorganized to make it easier to use for courses that cover only a subset of the material, and offers additional material including five new chapters. Students are introduced to Python and the basics of programming in the context of such computational concepts and techniques as exhaustive enumeration, bisection search, and efficient approximation algorithms. Although it covers such traditional topics as computational complexity and simple algorithms, the book focuses on a wide range of topics not found in most introductory texts, including information visualization, simulations to model randomness, computational techniques to understand data, and statistical techniques that inform (and misinform) as well as two related but relatively advanced topics: optimization problems and dynamic programming. This edition offers expanded material on statistics and machine learning and new chapters on Frequentist and Bayesian statistics.
Human computation is a new and evolving research area that centers around harnessing human intelligence to solve computational problems that are beyond the scope of existing Artificial Intelligence (AI) algorithms. With the growth of the Web, human computation systems can now leverage the abilities of an unprecedented number of people via the Web to perform complex computation. There are various genres of human computation applications that exist today. Games with a purpose (e.g., the ESP Game) specifically target online gamers who generate useful data (e.g., image tags) while playing an enjoyable game. Crowdsourcing marketplaces (e.g., Amazon Mechanical Turk) are human computation systems that coordinate workers to perform tasks in exchange for monetary rewards. In identity verification tasks, users perform computation in order to gain access to some online content; an example is reCAPTCHA, which leverages millions of users who solve CAPTCHAs every day to correct words in books that optical character recognition (OCR) programs fail to recognize with certainty. This book is aimed at achieving four goals: (1) defining human computation as a research area; (2) providing a comprehensive review of existing work; (3) drawing connections to a wide variety of disciplines, including AI, Machine Learning, HCI, Mechanism/Market Design and Psychology, and capturing their unique perspectives on the core research questions in human computation; and (4) suggesting promising research directions for the future. Table of Contents: Introduction / Human Computation Algorithms / Aggregating Outputs / Task Routing / Understanding Workers and Requesters / The Art of Asking Questions / The Future of Human Computation
When, in 1984?86, Richard P. Feynman gave his famous course on computation at the California Institute of Technology, he asked Tony Hey to adapt his lecture notes into a book. Although led by Feynman, the course also featured, as occasional guest speakers, some of the most brilliant men in science at that time, including Marvin Minsky, Charles Bennett, and John Hopfield. Although the lectures are now thirteen years old, most of the material is timeless and presents a ?Feynmanesque? overview of many standard and some not-so-standard topics in computer science such as reversible logic gates and quantum computers.
Perspectives in Computation covers three broad topics: the computation process & its limitations; the search for computational efficiency; & the role of quantum mechanics in computation.
The last lecture course that Nobel Prize winner Richard P. Feynman gave to students at Caltech from 1983 to 1986 was not on physics but on computer science. The first edition of the Feynman Lectures on Computation, published in 1996, provided an overview of standard and not-so-standard topics in computer science given in Feynman’s inimitable style. Although now over 20 years old, most of the material is still relevant and interesting, and Feynman’s unique philosophy of learning and discovery shines through. For this new edition, Tony Hey has updated the lectures with an invited chapter from Professor John Preskill on “Quantum Computing 40 Years Later”. This contribution captures the progress made toward building a quantum computer since Feynman’s original suggestions in 1981. The last 25 years have also seen the “Moore’s law” roadmap for the IT industry coming to an end. To reflect this transition, John Shalf, Senior Scientist at Lawrence Berkeley National Laboratory, has contributed a chapter on “The Future of Computing beyond Moore’s Law”. The final update for this edition is an attempt to capture Feynman’s interest in artificial intelligence and artificial neural networks. Eric Mjolsness, now a Professor of Computer Science at the University of California Irvine, was a Teaching Assistant for Feynman’s original lecture course and his research interests are now the application of artificial intelligence and machine learning for multi-scale science. He has contributed a chapter called “Feynman on Artificial Intelligence and Machine Learning” that captures the early discussions with Feynman and also looks toward future developments. This exciting and important work provides key reading for students and scholars in the fields of computer science and computational physics.