[Nfbf-l] The Myth of Multitasking
Alan Dicey
adicey at bellsouth.net
Mon Nov 26 15:21:17 UTC 2012
The Myth of Multitasking
In one of the many letters he wrote to his son in the 1740s, Lord
Chesterfield offered the following advice: "There is time enough for
everything in the course of the day, if you do but one thing at once, but
there is not time enough in the year, if you will do two things at a
time." To Chesterfield, singular focus was not merely a practical way to
structure one's time; it was a mark of intelligence. "This steady and
undissipated attention to one object, is a sure mark of a superior genius;
as hurry, bustle, and agitation, are the never-failing symptoms of a weak
and frivolous mind."
In modern times, hurry, bustle, and agitation have become a regular way of
life for many people - so much so that we have embraced a word to describe
our efforts to respond to the many pressing demands on our time:
multitasking. Used for decades to describe the parallel processing
abilities of computers, multitasking is now shorthand for the human attempt
to do simultaneously as many things as possible, as quickly as possible,
preferably marshalling the power of as many technologies as possible.
In the late 1990s and early 2000s, one sensed a kind of exuberance about the
possibilities of multitasking. Advertisements for new electronic gadgets -
particularly the first generation of handheld digital devices - celebrated
the notion of using technology to accomplish several things at once. The
word multitasking began appearing in the "skills" sections of résumés, as
office workers restyled themselves as high-tech, high-performing team
players. "We have always multitasked - inability to walk and chew gum is a
time-honored cause for derision - but never so intensely or self-consciously
as now," James Gleick wrote in his 1999 book Faster. "We are multitasking
connoisseurs - experts in crowding, pressing, packing, and overlapping
distinct activities in our all-too-finite moments." An article in the New
York Times Magazine in 2001 asked, "Who can remember life before
multitasking? These days we all do it." The article offered advice on "How
to Multitask" with suggestions about giving your brain's "multitasking hot
spot" an appropriate workout.
But more recently, challenges to the ethos of multitasking have begun to
emerge. Numerous studies have shown the sometimes-fatal danger of using cell
phones and other electronic devices while driving, for example, and several
states have now made that particular form of multitasking illegal. In the
business world, where concerns about time-management are perennial, warnings
about workplace distractions spawned by a multitasking culture are on the
rise. In 2005, the BBC reported on a research study, funded by
Hewlett-Packard and conducted by the Institute of Psychiatry at the
University of London, that found, "Workers distracted by e-mail and phone
calls suffer a fall in IQ more than twice that found in marijuana
smokers." The psychologist who led the study called this new "infomania"
a serious threat to workplace productivity. One of the Harvard Business
Review's "Breakthrough Ideas" for 2007 was Linda Stone's notion of
"continuous partial attention," which might be understood as a subspecies of
multitasking: using mobile computing power and the Internet, we are
"constantly scanning for opportunities and staying on top of contacts,
events, and activities in an effort to miss nothing."
Dr. Edward Hallowell, a Massachusetts-based psychiatrist who specializes in
the treatment of attention deficit/hyperactivity disorder and has written a
book with the self-explanatory title CrazyBusy, has been offering therapies
to combat extreme multitasking for years; in his book he calls multitasking
a "mythical activity in which people believe they can perform two or more
tasks simultaneously." In a 2005 article, he described a new condition,
"Attention Deficit Trait," which he claims is rampant in the business world.
ADT is "purely a response to the hyperkinetic environment in which we live,"
writes Hallowell, and its hallmark symptoms mimic those of ADD. "Never in
history has the human brain been asked to track so many data points,"
Hallowell argues, and this challenge "can be controlled only by creatively
engineering one's environment and one's emotional and physical health."
Limiting multitasking is essential. Best-selling business advice author
Timothy Ferriss also extols the virtues of "single-tasking" in his book, The
4-Hour Workweek. Multitasking might also be taking a toll on the economy.
One study by researchers at the University of California at Irvine monitored
interruptions among office workers; they found that workers took an average
of twenty-five minutes to recover from interruptions such as phone calls or
answering e-mail and return to their original task. Discussing multitasking
with the New York Times in 2007, Jonathan B. Spira, an analyst at the
business research firm Basex, estimated that extreme multitasking -
information overload - costs the U.S. economy $650 billion a year in lost
productivity.
Changing Our Brains
To better understand the multitasking phenomenon, neurologists and
psychologists have studied the workings of the brain. In 1999, Jordan
Grafman, chief of cognitive neuroscience at the National Institute of
Neurological Disorders and Stroke (part of the National Institutes of
Health), used functional magnetic resonance imaging (fMRI) scans to
determine that when people engage in "task-switching" - that is,
multitasking behavior - the flow of blood increases to a region of the
frontal cortex called Brodmann area 10. (The flow of blood to particular
regions of the brain is taken as a proxy indication of activity in those
regions.) "This is presumably the last part of the brain to evolve, the most
mysterious and exciting part," Grafman told the New York Times in 2001 -
adding, with a touch of hyperbole, "It's what makes us most human." It is
also what makes multitasking a poor long-term strategy for learning. Other
studies, such as those performed by psychologist René Marois of Vanderbilt
University, have used fMRI to demonstrate the brain's response to handling
multiple tasks. Marois found evidence of a "response selection bottleneck"
that occurs when the brain is forced to respond to several stimuli at once.
As a result, task-switching leads to time lost as the brain determines which
task to perform. Psychologist David Meyer at the University of Michigan
believes that rather than a bottleneck in the brain, a process of "adaptive
executive control" takes place, which "schedules task processes
appropriately to obey instructions about their relative priorities and
serial order," as he described to the New Scientist. Unlike many other
researchers who study multitasking, Meyer is optimistic that, with
training, the brain can learn to task-switch more effectively, and there is
some evidence that certain simple tasks are amenable to such practice. But
his research has also found that multitasking contributes to the release of
stress hormones and adrenaline, which can cause long-term health problems if
not controlled, and contributes to the loss of short-term memory.
In one recent study, Russell Poldrack, a psychology professor at the
University of California, Los Angeles, found that "multitasking adversely
affects how you learn. Even if you learn while multitasking, that learning
is less flexible and more specialized, so you cannot retrieve the
information as easily." His research demonstrates that people use different
areas of the brain for learning and storing new information when they are
distracted: brain scans of people who are distracted or multitasking show
activity in the striatum, a region of the brain involved in learning new
skills; brain scans of people who are not distracted show activity in the
hippocampus, a region involved in storing and recalling information.
Discussing his research on National Public Radio recently, Poldrack warned,
"We have to be aware that there is a cost to the way that our society is
changing, that humans are not built to work this way. We're really built to
focus. And when we sort of force ourselves to multitask, we're driving
ourselves to perhaps be less efficient in the long run even though it
sometimes feels like we're being more efficient."
If, as Poldrack concluded, "multitasking changes the way people learn," what
might this mean for today's children and teens, raised with an excess of new
entertainment and educational technology, and avidly multitasking at a young
age? Poldrack calls this the "million-dollar question." Media multitasking -
that is, the simultaneous use of several different media, such as
television, the Internet, video games, text messages, telephones, and
e-mail - is clearly on the rise, as a 2006 report from the Kaiser Family
Foundation showed: in 1999, only 16 percent of the time people spent using
any of those media was spent on multiple media at once; by 2005, 26 percent
of media time was spent multitasking. "I multitask every single second I am
online," confessed one study participant. "At this very moment I am watching
TV, checking my e-mail every two minutes, reading a newsgroup about who shot
JFK, burning some music to a CD, and writing this message."
The Kaiser report noted several factors that increase the likelihood of
media multitasking, including "having a computer and being able to see a
television from it." Also, "sensation-seeking" personality types are more
likely to multitask, as are those living in "a highly TV-oriented
household." The picture that emerges of these pubescent multitasking
mavens is of a generation of great technical facility and intelligence but
of extreme impatience, unsatisfied with slowness and uncomfortable with
silence: "I get bored if it's not all going at once, because everything has
gaps - waiting for a website to come up, commercials on TV, etc." one
participant said. The report concludes on a very peculiar note, perhaps
intended to be optimistic: "In this media-heavy world, it is likely that
brains that are more adept at media multitasking will be passed along and
these changes will be naturally selected," the report states. "After all,
information is power, and if one can process more information all at once,
perhaps one can be more powerful." This is techno-social Darwinism, nature
red in pixel and claw.
Other experts aren't so sure. As neurologist Jordan Grafman told Time
magazine: "Kids that are instant messaging while doing homework, playing
games online and watching TV, I predict, aren't going to do well in the long
run." "I think this generation of kids is guinea pigs," educational
psychologist Jane Healy told the San Francisco Chronicle; she worries that
they might become adults who engage in "very quick but very shallow
thinking." Or, as the novelist Walter Kirn suggests in a deft essay in The
Atlantic, we might be headed for an "Attention-Deficit Recession."
Paying Attention
When we talk about multitasking, we are really talking about attention: the
art of paying attention, the ability to shift our attention, and, more
broadly, to exercise judgment about what objects are worthy of our
attention. People who have achieved great things often credit for their
success a finely honed skill for paying attention. When asked about his
particular genius, Isaac Newton responded that if he had made any
discoveries, it was "owing more to patient attention than to any other
talent."
William James, the great psychologist, wrote at length about the varieties
of human attention. In The Principles of Psychology (1890), he outlined the
differences among "sensorial attention," "intellectual attention," "passive
attention," and the like, and noted the "gray chaotic indiscriminateness" of
the minds of people who were incapable of paying attention. James compared
our stream of thought to a river, and his observations presaged the
cognitive "bottlenecks" described later by neurologists: "On the whole easy
simple flowing predominates in it, the drift of things is with the pull of
gravity, and effortless attention is the rule," he wrote. "But at intervals
an obstruction, a set-back, a log-jam occurs, stops the current, creates an
eddy, and makes things temporarily move the other way."
To James, steady attention was thus the default condition of a mature mind,
an ordinary state undone only by perturbation. To readers a century later,
that placid portrayal may seem alien - as though depicting a bygone world.
Instead, today's multitasking adult may find something more familiar in
James's description of the youthful mind: an "extreme mobility of the
attention" that "makes the child seem to belong less to himself than to
every object which happens to catch his notice." For some people, James
noted, this challenge is never overcome; such people only get their work
done "in the interstices of their mind-wandering." Like Chesterfield, James
believed that the transition from youthful distraction to mature attention
was in large part the result of personal mastery and discipline - and so
was illustrative of character. "The faculty of voluntarily bringing back a
wandering attention, over and over again," he wrote, "is the very root of
judgment, character, and will."
Today, our collective will to pay attention seems fairly weak. We require
advice books to teach us how to avoid distraction. In the not-too-distant
future we may even employ new devices to help us overcome the unintended
attention deficits created by today's gadgets. As one New York Times article
recently suggested, "Further research could help create clever technology,
like sensors or smart software that workers could instruct with their
preferences and priorities to serve as a high tech 'time nanny' to ease the
modern multitasker's plight." Perhaps we will all accept as a matter of
course a computer governor - like the devices placed on engines so that
people can't drive cars beyond a certain speed. Our technological governors
might prompt us with reminders to set mental limits when we try to do too
much, too quickly, all at once.
Then again, perhaps we will simply adjust and come to accept what James
called "acquired inattention." E-mails pouring in, cell phones ringing,
televisions blaring, podcasts streaming - all this may become background
noise, like the "din of a foundry or factory" that James observed workers
could scarcely avoid at first, but which eventually became just another
part of their daily routine. For the younger generation of multitaskers, the
great electronic din is an expected part of everyday life. And given what
neuroscience and anecdotal evidence have shown us, this state of constant
intentional self-distraction could well be of profound detriment to
individual and cultural well-being. When people do their work only in the
"interstices of their mind-wandering," with crumbs of attention rationed out
among many competing tasks, their culture may gain in information, but it
will surely weaken in wisdom.
--------------------------------------------------------------------------------Christine Rosen is a senior editor of The New Atlantis and a fellow at theEthics and Public Policy Center.Christine Rosen, "The Myth of Multitasking," The New Atlantis, Number 20, Spring 2008, pp. 105-110.
More information about the NFBF-L
mailing list