Perspectives in Computation covers three broad topics: the computation process & its limitations; the search for computational efficiency; & the role of quantum mechanics in computation.
By paying close attention to the metaphors of artificial intelligence and their consequences for the field's patterns of success and failure, this text argues for a reorientation of the field away from thought and toward activity. It offers a critical reconstruction of AI research.
Analyzes cognitive, social and technical issues of end user programming. Drawing on empirical research on existing end user systems, this text examines the importance of task-specific programming languages, visual application frameworks and collaborative work practices for end user computing.
If you look around you will find that all computer systems, from your portable devices to the strongest supercomputers, are heterogeneous in nature. The most obvious heterogeneity is the existence of computing nodes of different capabilities (e.g. multicore, GPUs, FPGAs, ...). But there are also other heterogeneity factors that exist in computing systems, like the memory system components, interconnection, etc. The main reason for these different types of heterogeneity is to have good performance with power efficiency. Heterogeneous computing results in both challenges and opportunities. This book discusses both. It shows that we need to deal with these challenges at all levels of the computing stack: from algorithms all the way to process technology. We discuss the topic of heterogeneous computing from different angles: hardware challenges, current hardware state-of-the-art, software issues, how to make the best use of the current heterogeneous systems, and what lies ahead. The aim of this book is to introduce the big picture of heterogeneous computing. Whether you are a hardware designer or a software developer, you need to know how the pieces of the puzzle fit together. The main goal is to bring researchers and engineers to the forefront of the research frontier in the new era that started a few years ago and is expected to continue for decades. We believe that academics, researchers, practitioners, and students will benefit from this book and will be prepared to tackle the big wave of heterogeneous computing that is here to stay.
Since the summer of 1973, when I became a Burroughs Research Fellow, my life has been very different from what it had been before. The daily routine changed: instead of going to the University each day, where I used to spend most of my time in the company of others, I now went there only one day a week and was most of the time -that is, when not travelling!- alone in my study. In my solitude, mail and the written word in general became more and more important. The circumstance that my employer and I had the Atlantic Ocean between us was a further incentive to keep a fairly complete record of what I was doing. The public part of that output found its place in what became known as "the EWD series", which can be viewed as a form of scientific correspondence, possible since the advent of the copier. (That same copier makes it hard to estimate its actual distribution: I myself made about two dozen copies of my texts, but their recipients were welcome to act as further nodes of the distribution tree. ) The decision to publish a se1ection from the EWD series in book form was at first highly embarrassing, but as the months went by I got used to the idea. As soon as some guiding principles had been adopted -preferably not published elsewhere, as varied and as representative as possible, etc.
This textbook is intended as a textbook for one-semester, introductory computer science courses aimed at undergraduate students from all disciplines. Self-contained and with no prerequisites, it focuses on elementary knowledge and thinking models. The content has been tested in university classrooms for over six years, and has been used in summer schools to train university and high-school teachers on teaching introductory computer science courses using computational thinking. This book introduces computer science from a computational thinking perspective. In computer science the way of thinking is characterized by three external and eight internal features, including automatic execution, bit-accuracy and abstraction. The book is divided into chapters on logic thinking, algorithmic thinking, systems thinking, and network thinking. It also covers societal impact and responsible computing material – from ICT industry to digital economy, from the wonder of exponentiation to wonder of cyberspace, and from code of conduct to best practices for independent work. The book’s structure encourages active, hands-on learning using the pedagogic tool Bloom's taxonomy to create computational solutions to over 200 problems of varying difficulty. Students solve problems using a combination of thought experiment, programming, and written methods. Only 300 lines of code in total are required to solve most programming problems in this book.
Smart Data: State-of-the-Art Perspectives in Computing and Applications explores smart data computing techniques to provide intelligent decision making and prediction services support for business, science, and engineering. It also examines the latest research trends in fields related to smart data computing and applications, including new computing theories, data mining and machine learning techniques. The book features contributions from leading experts and covers cutting-edge topics such as smart data and cloud computing, AI for networking, smart data deep learning, Big Data capture and representation, AI for Big Data applications, and more. Features Presents state-of-the-art research in big data and smart computing Provides a broad coverage of topics in data science and machine learning Combines computing methods with domain knowledge and a focus on applications in science, engineering, and business Covers data security and privacy, including AI techniques Includes contributions from leading researchers
The aim of the book is to familiarize the new generation of PhD students and postdoctoral fellows with the principles and methods of modern lattice field theory, which aims to resolve fundamental, non-perturbative questions about QCD without uncontrolled approximations.