A leading artificial intelligence researcher lays out a new approach to AI that will enable people to coexist successfully with increasingly intelligent machines.
"The most important book on AI this year." --The Guardian "Mr. Russell's exciting book goes deep, while sparkling with dry witticisms." --The Wall Street Journal "The most important book I have read in quite some time" (Daniel Kahneman); "A must-read" (Max Tegmark); "The book we've all been waiting for" (Sam Harris) A leading artificial intelligence researcher lays out a new approach to AI that will enable us to coexist successfully with increasingly intelligent machines In the popular imagination, superhuman artificial intelligence is an approaching tidal wave that threatens not just jobs and human relationships, but civilization itself. Conflict between humans and machines is seen as inevitable and its outcome all too predictable. In this groundbreaking book, distinguished AI researcher Stuart Russell argues that this scenario can be avoided, but only if we rethink AI from the ground up. Russell begins by exploring the idea of intelligence in humans and in machines. He describes the near-term benefits we can expect, from intelligent personal assistants to vastly accelerated scientific research, and outlines the AI breakthroughs that still have to happen before we reach superhuman AI. He also spells out the ways humans are already finding to misuse AI, from lethal autonomous weapons to viral sabotage. If the predicted breakthroughs occur and superhuman AI emerges, we will have created entities far more powerful than ourselves. How can we ensure they never, ever, have power over us? Russell suggests that we can rebuild AI on a new foundation, according to which machines are designed to be inherently uncertain about the human preferences they are required to satisfy. Such machines would be humble, altruistic, and committed to pursue our objectives, not theirs. This new foundation would allow us to create machines that are provably deferential and provably beneficial.
Like Mooki, the hero of Spike Lee's film Do the Right Thing artificially, intelligent systems have a hard time knowing what to do in all circumstances. Classical theories of perfect rationality prescribe the right thing for any occasion, but no finite agent can compute their prescriptions fast enough. In Do the Right Thing, the authors argue that a new theoretical foundation for artificial intelligence can be constructed in which rationality is a property of programs within a finite architecture, and their behaviour over time in the task environment, rather than a property of individual decisions.
New York Times Best Seller How will Artificial Intelligence affect crime, war, justice, jobs, society and our very sense of being human? The rise of AI has the potential to transform our future more than any other technology—and there’s nobody better qualified or situated to explore that future than Max Tegmark, an MIT professor who’s helped mainstream research on how to keep AI beneficial. How can we grow our prosperity through automation without leaving people lacking income or purpose? What career advice should we give today’s kids? How can we make future AI systems more robust, so that they do what we want without crashing, malfunctioning or getting hacked? Should we fear an arms race in lethal autonomous weapons? Will machines eventually outsmart us at all tasks, replacing humans on the job market and perhaps altogether? Will AI help life flourish like never before or give us more power than we can handle? What sort of future do you want? This book empowers you to join what may be the most important conversation of our time. It doesn’t shy away from the full range of viewpoints or from the most controversial issues—from superintelligence to meaning, consciousness and the ultimate physical limits on life in the cosmos.
A jaw-dropping exploration of everything that goes wrong when we build AI systems and the movement to fix them. Today’s “machine-learning” systems, trained by data, are so effective that we’ve invited them to see and hear for us—and to make decisions on our behalf. But alarm bells are ringing. Recent years have seen an eruption of concern as the field of machine learning advances. When the systems we attempt to teach will not, in the end, do what we want or what we expect, ethical and potentially existential risks emerge. Researchers call this the alignment problem. Systems cull résumés until, years later, we discover that they have inherent gender biases. Algorithms decide bail and parole—and appear to assess Black and White defendants differently. We can no longer assume that our mortgage application, or even our medical tests, will be seen by human eyes. And as autonomous vehicles share our streets, we are increasingly putting our lives in their hands. The mathematical and computational models driving these changes range in complexity from something that can fit on a spreadsheet to a complex system that might credibly be called “artificial intelligence.” They are steadily replacing both human judgment and explicitly programmed software. In best-selling author Brian Christian’s riveting account, we meet the alignment problem’s “first-responders,” and learn their ambitious plan to solve it before our hands are completely off the wheel. In a masterful blend of history and on-the ground reporting, Christian traces the explosive growth in the field of machine learning and surveys its current, sprawling frontier. Readers encounter a discipline finding its legs amid exhilarating and sometimes terrifying progress. Whether they—and we—succeed or fail in solving the alignment problem will be a defining human story. The Alignment Problem offers an unflinching reckoning with humanity’s biases and blind spots, our own unstated assumptions and often contradictory goals. A dazzlingly interdisciplinary work, it takes a hard look not only at our technology but at our culture—and finds a story by turns harrowing and hopeful.
Porchlight’s Management and Workplace Culture Book of The Year “[A] thoroughly fascinating exploration of the long interplay between power and the technologies of communication.” —Adam Frank, NPR Team Human is a manifesto—a fiery distillation of preeminent digital theorist Douglas Rushkoff’s most urgent thoughts on civilization and human nature. In one hundred lean and incisive statements, he argues that we are essentially social creatures, and that we achieve our greatest aspirations when we work together—not as individuals. Yet today society is threatened by a vast antihuman infrastructure that undermines our ability to connect. Money, once a means of exchange, is now a means of exploitation; education, conceived as way to elevate the working class, has become another assembly line; and the internet has only further divided us into increasingly atomized and radicalized groups. Team Human delivers a call to arms. If we are to resist and survive these destructive forces, we must recognize that being human is a team sport. In Rushkoff’s own words: “Being social may be the whole point.” Harnessing wide-ranging research on human evolution, biology, and psychology, Rushkoff shows that when we work together we realize greater happiness, productivity, and peace. If we can find the others who understand this fundamental truth and reassert our humanity—together—we can make the world a better place to be human.
A call-to-arms about the broken nature of artificial intelligence, and the powerful corporations that are turning the human-machine relationship on its head. We like to think that we are in control of the future of "artificial" intelligence. The reality, though, is that we -- the everyday people whose data powers AI -- aren't actually in control of anything. When, for example, we speak with Alexa, we contribute that data to a system we can't see and have no input into -- one largely free from regulation or oversight. The big nine corporations -- Amazon, Google, Facebook, Tencent, Baidu, Alibaba, Microsoft, IBM and Apple--are the new gods of AI and are short-changing our futures to reap immediate financial gain. In this book, Amy Webb reveals the pervasive, invisible ways in which the foundations of AI -- the people working on the system, their motivations, the technology itself -- is broken. Within our lifetimes, AI will, by design, begin to behave unpredictably, thinking and acting in ways which defy human logic. The big nine corporations may be inadvertently building and enabling vast arrays of intelligent systems that don't share our motivations, desires, or hopes for the future of humanity. Much more than a passionate, human-centered call-to-arms, this book delivers a strategy for changing course, and provides a path for liberating us from algorithmic decision-makers and powerful corporations.
"Humans may not be Earth's most intelligent beings for much longer: the world champions of chess, Go, and Jeopardy! are now all AIs. Given the rapid pace of progress in AI, many predict that it could advance to human-level intelligence within the next several decades. From there, it could quickly outpace human intelligence. What do these developments mean for the future of the mind? In Artificial You, Susan Schneider says that it is inevitable that AI will take intelligence in new directions, but urges that it is up to us to carve out a sensible path forward. As AI technology turns inward, reshaping the brain, as well as outward, potentially creating machine minds, it is crucial to beware. Homo sapiens, as mind designers, will be playing with "tools" they do not understand how to use: the self, the mind, and consciousness. Schneider argues that an insufficient grasp of the nature of these entities could undermine the use of AI and brain enhancement technology, bringing about the demise or suffering of conscious beings. To flourish, we must grasp the philosophical issues lying beneath the algorithms. At the heart of her exploration is a sober-minded discussion of what AI can truly achieve: Can robots really be conscious? Can we merge with AI, as tech leaders like Elon Musk and Ray Kurzweil suggest? Is the mind just a program? Examining these thorny issues, Schneider proposes ways we can test for machine consciousness, questions whether consciousness is an unavoidable byproduct of sophisticated intelligence, and considers the overall dangers of creating machine minds."--Provided by publisher.
Elon Musk named Our Final Invention one of five books everyone should read about the future—a Huffington Post Definitive Tech Book of 2013. Artificial Intelligence helps choose what books you buy, what movies you see, and even who you date. It puts the “smart” in your smartphone and soon it will drive your car. It makes most of the trades on Wall Street, and controls vital energy, water, and transportation infrastructure. But Artificial Intelligence can also threaten our existence. In as little as a decade, AI could match and then surpass human intelligence. Corporations and government agencies are pouring billions into achieving AI’s Holy Grail—human-level intelligence. Once AI has attained it, scientists argue, it will have survival drives much like our own. We may be forced to compete with a rival more cunning, more powerful, and more alien than we can imagine. Through profiles of tech visionaries, industry watchdogs, and groundbreaking AI systems, Our Final Invention explores the perils of the heedless pursuit of advanced AI. Until now, human intelligence has had no rival. Can we coexist with beings whose intelligence dwarfs our own? And will they allow us to? “If you read just one book that makes you confront scary high-tech realities that we’ll soon have no choice but to address, make it this one.” —The Washington Post “Science fiction has long explored the implications of humanlike machines (think of Asimov’s I, Robot), but Barrat’s thoughtful treatment adds a dose of reality.” —Science News “A dark new book . . . lays out a strong case for why we should be at least a little worried.” —The New Yorker