Полная версия
Smarter Than You Think: How Technology is Changing Our Minds for the Better
Copyright
William Collins
An imprint of HarperCollinsPublishers Ltd
77–85 Fulham Palace Road,
Hammersmith, London W6 8JB
WilliamCollinsBooks.com
First published in Great Britain by William Collins in 2013
Copyright © Clive Thompson 2013
Clive Thompson asserts the moral right to be identified as the author of this work
A catalogue record for this book is available from the British Library
All rights reserved under International and Pan-American Copyright Conventions. By payment of the required fees, you have been granted the non-exclusive, non-transferable right to access and read the text of this e-book on-screen. No part of this text may be reproduced, transmitted, down-loaded, decompiled, reverse engineered, or stored in or introduced into any information storage and retrieval system, in any form or by any means, whether electronic or mechanical, now known or hereinafter invented, without the express written permission of HarperCollins.
Source ISBN: 9780007427796
Ebook Edition © September 2013 ISBN: 9780007427789
Version: 2014-09-06
From the reviews of Smarter Than You Think:
‘We should be grateful to have such a clear-eyed and lucid interpreter of our changing technological culture as Clive Thompson. Smarter Than You Think is an important, insightful book about who we are, and who we are becoming’
Joshua Foer, New York Times bestselling author of Moonwalking with Einstein
‘Almost without noticing it, the internet has become our intellectual exoskeleton. Rather than just observing this evolution, Clive Thompson takes us to the people, places and technologies driving it, bringing deep reporting, storytelling and analysis to one of the most profound shifts in human history’
Chris Anderson, author of The Long Tail
‘There’s good news in this dazzling book: technology is not the enemy. Smarter Than You Think reports on how the digital world has helped individuals harness a powerful, collaborative intelligence – becoming better problem-solvers and more creative human beings’
Jane McGonigal, author of Reality is Broken
‘Thompson has started an important debate in this lively and accessible book’
Scotsman
Dedication
To Emily, Gabriel, and Zev
Contents
Cover
Title Page
Copyright
Praise
Dedication
The Rise of the Centaurs
We, the Memorious
Public Thinking
The New Literacies
The Art of Finding
The Puzzle-Hungry World
Digital School
Ambient Awareness
The Connected Society
Epilogue
Notes
Index
Acknowledgments
About the Author
About the Publisher
The Rise of the Centaurs_
Who’s better at chess—computers or humans?
The question has long fascinated observers, perhaps because chess seems like the ultimate display of human thought: the players sit like Rodin’s Thinker, silent, brows furrowed, making lightning-fast calculations. It’s the quintessential cognitive activity, logic as an extreme sport.
So the idea of a machine outplaying a human has always provoked both excitement and dread. In the eighteenth century, Wolfgang von Kempelen caused a stir1 with his clockwork Mechanical Turk—an automaton that played an eerily good game of chess, even beating Napoleon Bonaparte. The spectacle was so unsettling that onlookers cried out in astonishment when the Turk’s gears first clicked into motion. But the gears, and the machine, were fake; in reality, the automaton was controlled by a chess savant cunningly tucked inside the wooden cabinet. In 1915, a Spanish inventor unveiled a genuine, honest-to-goodness robot2 that could actually play chess—a simple endgame involving only three pieces, anyway. A writer for Scientific American fretted that the inventor “Would Substitute Machinery for the Human Mind.”
Eighty years later, in 1997, this intellectual standoff clanked to a dismal conclusion when world champion Garry Kasparov was defeated by IBM’s Deep Blue supercomputer in a tournament of six games. Faced with a machine that could calculate two hundred million positions a second3, even Kasparov’s notoriously aggressive and nimble style broke down. In its final game, Deep Blue used such a clever ploy—tricking Kasparov into letting the computer sacrifice a knight—that it trounced him in nineteen moves. “I lost my fighting spirit,”4 Kasparov said afterward, pronouncing himself “emptied completely.”5 Riveted, the journalists announced a winner. The cover of Newsweek proclaimed the event “The Brain’s Last Stand.”6 Doomsayers predicted that chess itself was over7. If machines could outthink even Kasparov, why would the game remain interesting? Why would anyone bother playing? What’s the challenge?
Then Kasparov did something unexpected8.
The truth is, Kasparov wasn’t completely surprised by Deep Blue’s victory. Chess grand masters had predicted for years9 that computers would eventually beat humans, because they understood the different ways humans and computers play. Human chess players learn by spending years studying10 the world’s best opening moves and endgames; they play thousands of games, slowly amassing a capacious, in-brain library of which strategies triumphed and which flopped. They analyze their opponents’ strengths and weaknesses, as well as their moods. When they look at the board, that knowledge manifests as intuition—a eureka moment when they suddenly spy the best possible move.
In contrast, a chess-playing computer has no intuition at all. It analyzes the game using brute force; it inspects the pieces currently on the board, then calculates all options. It prunes away moves that lead to losing positions, then takes the promising ones and runs the calculations again. After doing this a few times—and looking five or seven moves out—it arrives at a few powerful plays. The machine’s way of “thinking” is fundamentally unhuman. Humans don’t sit around crunching every possible move, because our brains can’t hold that much information at once. If you go eight moves out in a game of chess,11 there are more possible games than there are stars in our galaxy. If you total up every game possible? It outnumbers the atoms in the known universe. Ask chess grand masters, “How many moves can you see out?” and they’ll likely deliver the answer attributed to the Cuban grand master José Raúl Capablanca: “One, the best one.”12
The fight between computers and humans in chess was, as Kasparov knew, ultimately about speed. Once computers could see all games roughly seven moves out, they would wear humans down. A person might make a mistake; the computer wouldn’t. Brute force wins. As he pondered Deep Blue, Kasparov mused on these different cognitive approaches.
It gave him an audacious idea. What would happen if, instead of competing against one another, humans and computers collaborated? What if they played on teams together—one computer and a human facing off against another human and a computer? That way, he theorized, each might benefit from the other’s peculiar powers. The computer would bring the lightning-fast—if uncreative—ability to analyze zillions of moves, while the human would bring intuition and insight, the ability to read opponents and psych them out. Together, they would form what chess players later called a centaur: a hybrid beast endowed with the strengths of each.
In June 1998, Kasparov played the first public game of human-computer collaborative chess, which he dubbed “advanced chess,” against Veselin Topalov, a top-rated grand master. Each used a regular computer with off-the-shelf chess software and databases of hundreds of thousands of chess games, including some of the best ever played. They considered what moves the computer recommended; they examined historical databases to see if anyone had ever been in a situation like theirs before. Then they used that information to help plan. Each game was limited to sixty minutes, so they didn’t have infinite time to consult the machines; they had to work swiftly.
Kasparov found the experience “as disturbing as it was exciting.” Freed from the need to rely exclusively on his memory, he was able to focus more on the creative texture of his play. It was, he realized, like learning to be a race-car driver: He had to learn how to drive the computer, as it were—developing a split-second sense of which strategy to enter into the computer for assessment, when to stop an unpromising line of inquiry, and when to accept or ignore the computer’s advice. “Just as a good Formula One driver really knows his own car, so did we have to learn the way the computer program worked,” he later wrote. Topalov, as it turns out, appeared to be an even better Formula One “thinker” than Kasparov. On purely human terms, Kasparov was a stronger player; a month before, he’d trounced Topalov 4–0. But the centaur play evened the odds. This time, Topalov fought Kasparov to a 3–3 draw.13
In 2005, there was a “freestyle” chess tournament14 in which a team could consist of any number of humans or computers, in any combination. Many teams consisted of chess grand masters who’d won plenty of regular, human-only tournaments, achieving chess scores of 2,500 (out of 3,000). But the winning team didn’t include any grand masters at all. It consisted of two young New England men, Steven Cramton and Zackary Stephen (who were comparative amateurs, with chess rankings down around 1,400 to 1,700), and their computers.
Why could these relative amateurs beat chess players with far more experience and raw talent? Because Cramton and Stephen were expert at collaborating with computers. They knew when to rely on human smarts and when to rely on the machine’s advice. Working at rapid speed—these games, too, were limited to sixty minutes—they would brainstorm moves, then check to see what the computer thought, while also scouring databases to see if the strategy had occurred in previous games. They used three different computers simultaneously, running five different pieces of software; that way they could cross-check whether different programs agreed on the same move. But they wouldn’t simply accept what the machine accepted, nor would they merely mimic old games. They selected moves that were low-rated by the computer if they thought they would rattle their opponents psychologically.
In essence, a new form of chess intelligence was emerging. You could rank the teams like this: (1) a chess grand master was good; (2) a chess grand master playing with a laptop was better. But even that laptop-equipped grand master could be beaten by (3) relative newbies, if the amateurs were extremely skilled at integrating machine assistance. “Human strategic guidance combined with the tactical acuity of a computer,” Kasparov concluded, “was overwhelming.”
Better yet, it turned out these smart amateurs could even outplay a supercomputer on the level of Deep Blue. One of the entrants that Cramton and Stephen trounced in the freestyle chess tournament was a version of Hydra, the most powerful chess computer in existence15 at the time; indeed, it was probably faster and stronger than Deep Blue itself. Hydra’s owners let it play entirely by itself, using raw logic and speed to fight its opponents. A few days after the advanced chess event, Hydra destroyed the world’s seventh-ranked grand master in a man-versus-machine chess tournament.
But Cramton and Stephen beat Hydra. They did it using their own talents and regular Dell and Hewlett-Packard computers, of the type you probably had sitting on your desk in 2005, with software you could buy for sixty dollars.16 All of which brings us back to our original question here: Which is smarter at chess—humans or computers?
Neither.
It’s the two together, working side by side.
We’re all playing advanced chess these days. We just haven’t learned to appreciate it.
Our tools are everywhere, linked with our minds, working in tandem. Search engines answer our most obscure questions; status updates give us an ESP-like awareness of those around us; online collaborations let far-flung collaborators tackle problems too tangled for any individual. We’re becoming less like Rodin’s Thinker and more like Kasparov’s centaurs. This transformation is rippling through every part of our cognition—how we learn, how we remember, and how we act upon that knowledge emotionally, intellectually, and politically. As with Cramton and Stephen, these tools can make even the amateurs among us radically smarter than we’d be on our own, assuming (and this is a big assumption) we understand how they work. At their best, today’s digital tools help us see more, retain more, communicate more. At their worst, they leave us prey to the manipulation of the toolmakers. But on balance, I’d argue, what is happening is deeply positive. This book is about the transformation.
In a sense, this is an ancient story. The “extended mind” theory of cognition argues that the reason humans are so intellectually dominant is that we’ve always outsourced bits of cognition, using tools to scaffold our thinking into ever-more-rarefied realms. Printed books amplified our memory. Inexpensive paper and reliable pens made it possible to externalize our thoughts quickly. Studies show that our eyes zip around the page while performing long division on paper, using the handwritten digits as a form of prosthetic short-term memory.17 “These resources enable us to pursue18 manipulations and juxtapositions of ideas and data that would quickly baffle the un-augmented brain,” as Andy Clark, a philosopher of the extended mind, writes.
Granted, it can be unsettling to realize how much thinking already happens outside our skulls. Culturally, we revere the Rodin ideal—the belief that genius breakthroughs come from our gray matter alone. The physicist Richard Feynman once got into an argument about this with the historian Charles Weiner. Feynman understood the extended mind; he knew that writing his equations and ideas on paper was crucial to his thought. But when Weiner looked over a pile of Feynman’s notebooks, he called them a wonderful “record of his day-to-day work.” No, no, Feynman replied testily. They weren’t a record of his thinking process. They were his thinking process:
“I actually did the work on the paper,”19 he said.
“Well,” Weiner said, “the work was done in your head, but the record of it is still here.”
“No, it’s not a record, not really. It’s working. You have to work on paper and this is the paper. Okay?”
Every new tool shapes the way we think, as well as what we think about. The printed word helped make our cognition linear and abstract,20 along with vastly enlarging our stores of knowledge. Newspapers shrank the world; then the telegraph shrank it even more dramatically. With every innovation, cultural prophets bickered over whether we were facing a technological apocalypse or a utopia. Depending on which Victorian-age pundit you asked, the telegraph was either going usher in an era of world peace (“It is impossible that old prejudices and hostilities should longer exist,”21 as Charles F. Briggs and Augustus Maverick intoned) or drown us in a Sargasso of idiotic trivia (“We are eager to tunnel under the Atlantic22 … but perchance the first news that will leak through into the broad, flapping American ear will be that the Princess Adelaide has the whooping cough,” as Thoreau opined). Neither prediction was quite right, of course, yet neither was quite wrong. The one thing that both apocalyptics and utopians understand and agree upon is that every new technology pushes us toward new forms of behavior while nudging us away from older, familiar ones. Harold Innis—the lesser-known but arguably more interesting intellectual midwife of Marshall McLuhan—called this the bias of a new tool.23 Living with new technologies means understanding how they bias everyday life.
What are the central biases of today’s digital tools? There are many, but I see three big ones that have a huge impact on our cognition. First, they allow for prodigious external memory: smartphones, hard drives, cameras, and sensors routinely record more information than any tool before them. We’re shifting from a stance of rarely recording our ideas and the events of our lives to doing it habitually. Second, today’s tools make it easier for us to find connections—between ideas, pictures, people, bits of news—that were previously invisible. Third, they encourage a superfluity of communication and publishing. This last feature has many surprising effects that are often ill understood. Any economist can tell you that when you suddenly increase the availability of a resource, people do more things with it, which also means they do increasingly unpredictable things. As electricity became cheap and ubiquitous in the West, its role expanded from things you’d expect—like nighttime lighting—to the unexpected and seemingly trivial: battery-driven toy trains, electric blenders, vibrators. The superfluity of communication today has produced everything from a rise in crowd-organized projects like Wikipedia to curious new forms of expression: television-show recaps, map-based storytelling, discussion threads that spin out of a photo posted to a smartphone app, Amazon product-review threads wittily hijacked for political satire. Now, none of these three digital biases is immutable, because they’re the product of software and hardware, and can easily be altered or ended if the architects of today’s tools (often corporate and governmental) decide to regulate the tools or find they’re not profitable enough. But right now, these big effects dominate our current and near-term landscape.
In one sense, these three shifts—infinite memory, dot connecting, explosive publishing—are screamingly obvious to anyone who’s ever used a computer. Yet they also somehow constantly surprise us by producing ever-new “tools for thought”24 (to use the writer Howard Rheingold’s lovely phrase) that upend our mental habits in ways we never expected and often don’t apprehend even as they take hold. Indeed, these phenomena have already woven themselves so deeply into the lives of people around the globe that it’s difficult to stand back and take account of how much things have changed and why. While this book maps out what I call the future of thought, it’s also frankly rooted in the present, because many parts of our future have already arrived, even if they are only dimly understood. As the sci-fi author William Gibson famously quipped: “The future is already here25—it’s just not very evenly distributed.” This is an attempt to understand what’s happening to us right now, the better to see where our augmented thought is headed. Rather than dwell in abstractions, like so many marketers and pundits—not to mention the creators of technology, who are often remarkably poor at predicting how people will use their tools—I focus more on the actual experiences of real people.
To provide a concrete example of what I’m talking about, let’s take a look at something simple and immediate: my activities while writing the pages you’ve just read.
As I was working, I often realized I couldn’t quite remember a detail and discovered that my notes were incomplete. So I’d zip over to a search engine. (Which chess piece did Deep Blue sacrifice when it beat Kasparov? The knight!) I also pushed some of my thinking out into the open: I blogged admiringly about the Spanish chess-playing robot from 1915, and within minutes commenters offered smart critiques. (One pointed out that the chess robot wasn’t that impressive because it was playing an endgame that was almost impossible to lose: the robot started with a rook and a king, while the human opponent had only a mere king.) While reading Kasparov’s book How Life Imitates Chess on my Kindle, I idly clicked on “popular highlights” to see what passages other readers had found interesting—and wound up becoming fascinated by a section on chess strategy I’d only lightly skimmed myself. To understand centaur play better, I read long, nuanced threads on chess-player discussion groups, effectively eavesdropping on conversations of people who know chess far better than I ever will. (Chess players who follow the new form of play seem divided—some think advanced chess is a grim sign of machines’ taking over the game, and others think it shows that the human mind is much more valuable than computer software.) I got into a long instant-messaging session with my wife, during which I realized that I’d explained the gist of advanced chess better than I had in my original draft, so I cut and pasted that explanation into my notes. As for the act of writing itself? Like most writers, I constantly have to fight the procrastinator’s urge to meander online, idly checking Twitter links and Wikipedia entries in a dreamy but pointless haze—until I look up in horror and realize I’ve lost two hours of work, a missing-time experience redolent of a UFO abduction. So I’d switch my word processor into full-screen mode, fading my computer desktop to black so I could see nothing but the page, giving me temporary mental peace.
In this book I explore each of these trends. First off, there’s the emergence of omnipresent computer storage, which is upending the way we remember, both as individuals and as a culture. Then there’s the advent of “public thinking”: the ability to broadcast our ideas and the catalytic effect that has both inside and outside our minds. We’re becoming more conversational thinkers—a shift that has been rocky, not least because everyday public thought uncorks the incivility and prejudices that are commonly repressed in face-to-face life. But at its best (which, I’d argue, is surprisingly often), it’s a thrilling development, reigniting ancient traditions of dialogue and debate. At the same time, there’s been an explosion of new forms of expression that were previously too expensive for everyday thought—like video, mapping, or data crunching. Our social awareness is shifting, too, as we develop ESP-like “ambient awareness,” a persistent sense of what others are doing and thinking. On a social level, this expands our ability to understand the people we care about. On a civic level, it helps dispel traditional political problems like “pluralistic ignorance,” catalyzing political action, as in the Arab Spring.
Are these changes good or bad for us? If you asked me twenty years ago, when I first started writing about technology, I’d have said “bad.” In the early 1990s, I believed that as people migrated online, society’s worst urges might be uncorked: pseudonymity would poison online conversation, gossip and trivia would dominate, and cultural standards would collapse. Certainly some of those predictions have come true, as anyone who’s wandered into an angry political forum knows. But the truth is, while I predicted the bad stuff, I didn’t foresee the good stuff. And what a torrent we have: Wikipedia, a global forest of eloquent bloggers, citizen journalism, political fact-checking—or even the way status-update tools like Twitter have produced a renaissance in witty, aphoristic, haiku-esque expression. If this book accentuates the positive, that’s in part because we’ve been so flooded with apocalyptic warnings of late. We need a new way to talk clearly about the rewards and pleasures of our digital experiences—one that’s rooted in our lived experience and also detangled from the hype of Silicon Valley.