CS3 In the News
Cirrus clouds form around mineral dust and metallic particles, study finds.

"The warmer climate gets, the faster the climate zones are shifting. This could make it harder for plants and animals to adjust," said lead author Irina Mahlstein.
The study is the first to look at the accelerating pace of the shifting of climate zones, which are areas of the Earth defined by annual and seasonal cycles of temperature and precipitation, as well as temperature and precipitation thresholds of plant species. Over 30 different climate zones are found on Earth; examples include the equatorial monsoonal zone, the polar tundra zone, and cold arid desert zone.
"A shift in the climate zone is probably a better measure of 'reality' for living systems, more so than changing temperature by a degree or precipitation by a centimeter," said Mahlstein.
The scientists used climate model simulations and a well-known ecosystem classification scheme to look at the shifts between climate zones over a two-century period, 1900 to 2098. The team found that for the initial 2 ° Celsius (3.6 ° Fahrenheit) of warming, about 5 percent of Earth's land area shifts to a new climate zone. The models show that the pace of change quickens for the next 2 ° Celsius of warming, and an additional 10 percent of the land area shifts to a new climate zone. "Pace of shifts in climate regions increases with global temperature" was published online in the journal Nature Climate Change on April 21.
Certain regions of the globe, such as northern middle and high latitudes, will undergo more changes than other regions, such as the tropics, the scientists found. In the tropics, mountainous regions will experience bigger changes than their surrounding low-altitude areas.
In the coming century, the findings suggest that frost climates–the coldest climate zone of the planet–are largely decreasing. Generally, dry regions in different areas of the globe are increasing, and a large fraction of land area is changing from cool summers to hot summers.
The scientists also investigated whether temperature or precipitation made the greater impact on how much of the land area changed zones. "We found that temperature is the main factor, at least through the end of this century," said Mahlstein.
This story is adapted from a news article at esrl.noaa.gov.
By Charles C. Mann
New technology and a little-known energy source suggest that fossil fuels may not be finite. This would be a miracle—and a nightmare.
As the great research ship Chikyu left Shimizu in January to mine the explosive ice beneath the Philippine Sea, chances are good that not one of the scientists aboard realized they might be closing the door on Winston Churchill’s world. Their lack of knowledge is unsurprising; beyond the ranks of petroleum-industry historians, Churchill’s outsize role in the history of energy is insufficiently appreciated.
Winston Leonard Spencer Churchill was appointed First Lord of the Admiralty in 1911. With characteristic vigor and verve, he set about modernizing the Royal Navy, jewel of the empire. The revamped fleet, he proclaimed, should be fueled with oil, rather than coal—a decision that continues to reverberate in the present. Burning a pound of fuel oil produces about twice as much energy as burning a pound of coal. Because of this greater energy density, oil could push ships faster and farther than coal could.
Churchill’s proposal led to emphatic dispute. The United Kingdom had lots of coal but next to no oil. At the time, the United States produced almost two-thirds of the world’s petroleum; Russia produced another fifth. Both were allies of Great Britain. Nonetheless, Whitehall was uneasy about the prospect of the Navy’s falling under the thumb of foreign entities, even if friendly. The solution, Churchill told Parliament in 1913, was for Britons to become “the owners, or at any rate, the controllers at the source of at least a proportion of the supply of natural oil which we require.” Spurred by the Admiralty, the U.K. soon bought 51 percent of what is now British Petroleum, which had rights to oil “at the source”: Iran (then known as Persia). The concessions’ terms were so unpopular in Iran that they helped spark a revolution. London worked to suppress it. Then, to prevent further disruptions, Britain enmeshed itself ever more deeply in the Middle East, working to install new shahs in Iran and carve Iraq out of the collapsing Ottoman Empire.
Churchill fired the starting gun, but all of the Western powers joined the race to control Middle Eastern oil. Britain clawed past France, Germany, and the Netherlands, only to be overtaken by the United States, which secured oil concessions in Turkey, Iraq, Bahrain, Kuwait, and Saudi Arabia. The struggle created a long-lasting intercontinental snarl of need and resentment. Even as oil-consuming nations intervened in the affairs of oil-producing nations, they seethed at their powerlessness; oil producers exacted huge sums from oil consumers but chafed at having to submit to them. Decades of turmoil—oil shocks in 1973 and 1979, failed programs for “energy independence,” two wars in Iraq—have left unchanged this fundamental, Churchillian dynamic, a toxic mash of anger and dependence that often seems as basic to global relations as the rotation of the sun.
All of this was called into question by the voyage of the Chikyu (“Earth”), a $540 million Japanese deep-sea drilling vessel that looks like a billionaire’s yacht with a 30-story oil derrick screwed into its back. The Chikyu, a floating barrage of superlatives, is the biggest, glitziest, most sophisticated research vessel ever constructed, and surely the only one with a landing pad for a 30-person helicopter. The central derrick houses an enormous floating drill with a six-mile “string” that has let the Chikyu delve deeper beneath the ocean floor than any other ship.
The Chikyu, which first set out in 2005, was initially intended to probe earthquake-generating zones in the planet’s mantle, a subject of obvious interest to seismically unstable Japan. Its present undertaking was, if possible, of even greater importance: trying to develop an energy source that could free not just Japan but much of the world from the dependence on Middle Eastern oil that has bedeviled politicians since Churchill’s day.
In the 1970s, geologists discovered crystalline natural gas—methane hydrate, in the jargon—beneath the seafloor. Stored mostly in broad, shallow layers on continental margins, methane hydrate exists in immense quantities; by some estimates, it is twice as abundant as all other fossil fuels combined. Despite its plenitude, gas hydrate was long subject to petroleum-industry skepticism. These deposits—water molecules laced into frigid cages that trap “guest molecules” of natural gas—are strikingly unlike conventional energy reserves. Ice you can set on fire! Who could take it seriously? But as petroleum prices soared, undersea-drilling technology improved, and geological surveys accumulated, interest rose around the world. The U.S. Department of Energy has been funding a methane-hydrate research program since 1982.
Nowhere has the interest been more serious than Japan. Unlike Britain and the United States, the Japanese failed to become “the owners, or at any rate, the controllers” of any significant amount of oil. (Not that Tokyo didn’t try: it bombed Pearl Harbor mainly to prevent the U.S. from blocking its attempted conquest of the oil-rich Dutch East Indies.) Today, Churchill’s nightmare has come true for Japan: it is a military and industrial power almost wholly dependent on foreign energy. It is the world’s third-biggest net importer of crude oil, the second-biggest importer of coal, and the biggest importer of liquefied natural gas. Not once has a Japanese politician expressed happiness at this state of affairs.
Japan’s methane-hydrate program began in 1995. Its scientists quickly focused on the Nankai Trough, about 200 miles southwest of Tokyo, an undersea earthquake zone where two pieces of the Earth’s crust jostle each other. Step by step, year by year, a state-owned enterprise now called the Japan Oil, Gas, and Metals National Corporation (JOGMEC) dug test wells, made measurements, and obtained samples of the hydrate deposits: 130-foot layers of sand and silt, loosely held together by methane-rich ice. The work was careful, slow, orderly, painstakingly analytical—the kind of process that seems intended to snuff out excited newspaper headlines. But it progressed with the same remorselessness that in the 1960s and ’70s had transformed offshore oil wells from Waterworld-style exoticisms to mainstays of the world economy.
In January, 18 years after the Japanese program began, the Chikyu left the Port of Shimizu, midway up the main island’s eastern coastline, to begin a “production” test—an attempt to harvest usefully large volumes of gas, rather than laboratory samples. Many questions remained to be answered, the project director, Koji Yamamoto, told me before the launch. JOGMEC hadn’t figured out the best way to mine hydrate, or how to ship the resultant natural gas to shore. Costs needed to be brought down. “It will not be ready for 10 years,” Yamamoto said. “But I believe it will be ready.” What would happen then, he allowed, would be “interesting.”
Already the petroleum industry has been convulsed by hydraulic fracturing, or “fracking”—a technique for shooting water mixed with sand and chemicals into rock, splitting it open, and releasing previously inaccessible oil, referred to as “tight oil.” Still more important, fracking releases natural gas, which, when yielded from shale, is known as shale gas. (Petroleum is a grab-bag term for all nonsolid hydrocarbon resources—oil of various types, natural gas, propane, oil precursors, and so on—that companies draw from beneath the Earth’s surface. The stuff that catches fire around stove burners is known by a more precise term, natural gas, referring to methane, a colorless, odorless gas that has the same chemical makeup no matter what the source—ordinary petroleum wells, shale beds, or methane hydrate.) Fracking has been attacked as an environmental menace to underground water supplies, and may eventually be greatly restricted. But it has also unleashed so much petroleum in North America that the International Energy Agency, a Paris-based consortium of energy-consuming nations, predicted in November that by 2035, the United States will become “all but self-sufficient in net terms.” If the Chikyu researchers are successful, methane hydrate could have similar effects in Japan. And not just in Japan: China, India, Korea, Taiwan, and Norway are looking to unlock these crystal cages, as are Canada and the United States.
Not everyone thinks JOGMEC will succeed. But methane hydrate is being developed in much the same methodical way that shale gas was developed before it, except by a bigger, more international group of researchers. Shale gas, too, was subject to skepticism wide and loud. The egg on naysayers’ faces suggests that it would be foolish to ignore the prospects for methane hydrate—and more foolish still not to consider the potential consequences.
If methane hydrate allows much of the world to switch from oil to gas, the conversion would undermine governments that depend on oil revenues, especially petro-autocracies like Russia, Iran, Venezuela, Iraq, Kuwait, and Saudi Arabia. Unless oil states are exceptionally well run, a gush of petroleum revenues can actually weaken their economies by crowding out other business. Worse, most oil nations are so corrupt that social scientists argue over whether there is an inherent bond—a “resource curse”—between big petroleum deposits and political malfeasance. It seems safe to say that few Americans would be upset if a plunge in demand eliminated these countries’ hold over the U.S. economy. But those same people might not relish the global instability—a belt of financial and political turmoil from Venezuela to Turkmenistan—that their collapse could well unleash.
On a broader level still, cheap, plentiful natural gas throws a wrench into efforts to combat climate change. Avoiding the worst effects of climate change, scientists increasingly believe, will require “a complete phase-out of carbon emissions … over 50 years,” in the words of one widely touted scientific estimate that appeared in January. A big, necessary step toward that goal is moving away from coal, still the second-most-important energy source worldwide. Natural gas burns so much cleaner than coal that converting power plants from coal to gas—a switch promoted by the deluge of gas from fracking—has already reduced U.S. greenhouse-gas emissions to their lowest levels since Newt Gingrich’s heyday.
Yet natural gas isn’t that clean; burning it produces carbon dioxide. Researchers view it as a temporary “bridge fuel,” something that can power nations while they make the transition away from oil and coal. But if societies do not take advantage of that bridge to enact anti-carbon policies, says Michael Levi, the director of the Program on Energy Security and Climate Change at the Council on Foreign Relations, natural gas could be “a bridge from the coal-fired past to the coal-fired future.”
“Methane hydrate could be a new energy revolution,” Christopher Knittel, a professor of energy economics at the Massachusetts Institute of Technology, told me. “It could help the world while we reduce greenhouse gases. Or it could undermine the economic rationale for investing in renewable, carbon-free energy around the world”—just as abundant shale gas from fracking has already begun to undermine it in the United States. “The one path is a boon. The other—I’ve used words like catastrophe.” He paused; I thought I detected a sigh. “I wouldn’t bet on us making the right decisions.”
A few years after I graduated from college, I drove with a friend to Southern California, a place I’d never been. I saw a little of Los Angeles, then went north and spent a few days bumbling through the San Joaquin Valley. Going about Bakersfield one night, I got hopelessly lost and ended up at a chain-link fence. Behind the fence were thousands of oil pumps, nodding up and down like so many giant plastic drinking birds. Enshrouding the pumps was a spiderweb of pipes and electrical wires, vast and complex beyond reason, lights and machinery stretching out across the desert farther than I could see. A giant, hypermodern petroleum operation barely 100 miles from Los Angeles! I couldn’t believe it. As I stood gawping, a policeman drove by. I asked him when this complex had sprung up. He looked at me like I was an idiot. “They’ve been drilling here since 1899,” he said.
I was standing by the Kern River oil field, one of the best-known petroleum deposits in the United States. Because I had somehow missed geology in school, I had been left with the vague idea that oil is found in big subterranean pools, like the underground lake where Voldemort conceals part of his soul in the Harry Potter series. In fact, petroleum is usually contained in solid sandstone or limestone strata, which are riddled, spongelike, with minute pores. Or it can occur in thin sheets between layers of shale. Looking at the nodding wells, I had the notion that they were drawing a uniform substance from the ground, a black liquid like the inky water in Voldemort’s lake. Instead, petroleum occurs as a crazy stew of different compounds: oil of various grades mixed with methane, ethane, propane, butane, and other hydrocarbons. Squashed into stone hundreds or thousands of feet underground, this jumble of liquid and gas is usually under great pressure. Layers, or “caps,” of impermeable rock prevent it from seeping to the surface. When drilling bores through the caps, petroleum shoots up in orthodox gusher fashion.
For a long time, companies collected oil and discarded the methane that burbled up with it, often by burning the gas in a cinematic flare atop special derricks, or even simply dumping it into the atmosphere. People did use natural gas for energy—gaslights have existed since the days of Jane Austen—but transporting it was costly. Unlike liquid oil, which could be poured into containers and carried on a railroad network that had already been built and paid for by somebody else, gaseous methane had to be pumped through sealed tubes to its destination, which required energy firms and utilities to lay thousands upon thousands of miles of pipeline. Not until the Second World War and war-production advances in welding did this effort gather speed. (Methane can be cooled into a liquid and transported in pressurized tanks that are loaded and unloaded in special facilities, but this is also expensive.) Oil from wells in Texas is readily dispatched via tanker to Europe or Asia, but even today, natural gas from the same wells is often effectively limited to use in the United States.
From the beginning, it was evident that the Kern River field was rich with oil, millions upon millions of barrels. (A barrel, the unit of oil measurement, is 42 gallons; depending on the grade, a ton of oil is six to eight barrels.) Wildcatters poured into the area, throwing up derricks, boring wells, and pulling out what they could. In 1949, after 50 years of drilling, analysts estimated that just 47 million barrels remained in reserves—a rounding error in the oil business. Kern River, it seemed, was nearly played out. Instead, oil companies removed 945 million barrels in the next 40 years. In 1989, analysts again estimated Kern reserves: 697 million barrels. By 2009, Kern had produced more than 1.3 billion additional barrels, and reserves were estimated to be almost 600 million barrels.
What does it mean when oil companies say they have so many million barrels in reserves? How much energy is in the ground? When will we begin running out? As the history of the Kern River field suggests, these questions are not easy to answer. Indeed, Ph.D.‑toting experts have bombarded Americans for half a century with totally contradictory responses. On one side, pessimists claim that the planet is slowly running out of petroleum. “Turn down the thermostat!” they cry. “Stuff insulation in your walls!” “Buy a hybrid!” “Conserve!” From the other side come equally loud shouts insisting that there are vast, untapped petroleum deposits in Alaska and Alberta and off the coast of Virginia, that geysers of natural gas exist in the shale beds of Pennsylvania and North Dakota, and that huge oil patches await extraction in the deep ocean. “Drill, baby, drill!” “The end of oil!” Al Gore or Sarah Palin, Cassandra or Pollyanna, which side is right? The back-and-forth would be comical if the stakes didn’t involve the fate of human civilization.
When gasoline supplies drop, TV news reporters like to wring their hands at the drivers mobbing the corner Exxon. But the motorists’ panic reflects a basic truth: economic growth and energy use have marched in lockstep for generations. Between 1900 and 2000, global energy consumption rose roughly 17-fold, the University of Manitoba environmental scientist Vaclav Smil has calculated, while economic output rose 16-fold—“as close a link as one may find in the unruly realm of economic affairs.” Petroleum has wreaked all kinds of social and environmental havoc, but a steady supply of oil and gas remains just as central to the world’s economic well-being as it was in Churchill’s day. According to the National Bureau of Economic Research, the United States has experienced 11 recessions since the end of the Second World War. All but one were associated with spikes in energy costs—specifically, abrupt jumps in the price of oil.
Understanding this dependence, the oil industry was shaken by a speech in 1956 by M. King Hubbert, a prominent geophysicist at Shell Oil. When a company moves into a field, it grabs the easy, cheap oil first. Tapping the rest gets progressively more difficult and expensive. Eventually, Hubbert observed, conditions get so tough that production levels off—it peaks. After the peak, decline is unstoppable, the fall as ineluctable as the rise. Hubbert used his theory to predict that the crude-oil yield in the continental United States would flatten between 1965 and 1970 (he didn’t include Alaska and most offshore oil areas). Coming at a time when estimates by the U.S. Geological Survey and the petroleum industry were constantly rising, this claim was derided; indeed, Hubbert claimed that just before giving his speech, a Shell official tried to get him to back off.
Hubbert, not the least self-confident of men, stood his ground, even after he left Shell and in 1964 went to work for the Geological Survey. Unluckily for him, his most prominent critic was now his boss: Vincent E. McKelvey, a long-serving geologist at USGS who would become its director in 1971. As the University of Iowa historian Tyler Priest has documented, McKelvey’s USGS issued a stream of optimistic assessments about the country’s oil future. So did its counterparts in the oil industry. Meanwhile, Hubbert cranked out papers taking the opposite stance, none of them published by the Geological Survey. Inevitably, the dispute grew personal. Three days after McKelvey became the USGS director, he took away Hubbert’s secretary, a harsh measure in the days before e‑mail. According to Priest, Hubbert ended up having to write all his correspondence in longhand; his wife typed his reports at home. Hubbert struck back by helping to kill McKelvey’s nominations to the National Academy of Sciences and the American Academy of Arts and Sciences.
In a blow to McKelvey, Hubbert’s prediction proved to be correct. As domestic crude-oil production peaked and then fell, former Interior Secretary Stewart Udall mocked the sunny claims from the Geological Survey as “an enormous energy balloon of inflated promises and boundless optimism [that] had long since lost touch with any mainland reality.” If Udall were reappointed Interior secretary, he said, “the first thing I would do would be to kick McKelvey out.” In 1977, newly elected President Jimmy Carter, a Hubbertian, forced McKelvey to resign—the first such ouster, Priest notes, “in the Survey’s 98-year history.”
Hubbert’s message of scarcity resonated at a time when the United States was haunted by the specter of Middle Eastern oil blockades. In a nationwide address, President Carter proclaimed that the planet’s proven oil reserves could be consumed “by the end of the next decade.” To forestall the disaster, he fired a volley of energy-efficiency measures: gas-mileage regulation, home-appliance energy standards, conservation tax credits, subsidies for insulation and weatherization. Congress enacted incentives and restrictions to induce industry to switch from supposedly scarce oil and natural gas to coal, which the U.S. has in abundance.
Alas, petroleum firms found so much crude oil in the 1980s that by the 1990s, prices (after adjusting for inflation) had fallen to one-fifth of what they had been during the Carter administration. Estimates of reserves rose and rose again. Energy conservation faltered; oil and gas were too cheap to be worth saving.
The argument has nonetheless continued, pessimists and optimists hammering at each other like Montagues and Capulets. Most of the Hubbertians are physical scientists; most of the McKelveyans, social scientists. Central to the conflict is their differing concepts of a reserve. Recall, as an example, the Kern River field. Its thousands of nodding pumps are siphoning up oil so thick and heavy that it almost doesn’t float on water. Although drillers knew from the first that the field was abundant, they could barely wrest any of this goop from the ground, a factor reflected in the first estimate of the reserve (47 million barrels of recoverable oil). Between that estimate and the second (697 million barrels), engineers developed a precursor to fracking: shooting hot steam down Kern River wells to thin the oil and force it out of the stone. At first, the process was hideously inefficient: heating the water to produce the steam required as much as 40 percent of the oil that came out of the wells. Burning unrefined crude oil released torrents of pollution: nitrous oxide, sulfur dioxide, carbon dioxide. But it squeezed out petroleum that had seemed impossible to reach.
At the same time, the industry learned how to burrow farther into the Earth, opening up previously inaccessible deposits. In 1998, an oil rig near the Kern River field drilled thousands of feet deeper than any previous attempt in the area. At 17,657 feet, the well blew out in a classic gusher. Flames shot 300 feet in the air. The blast destroyed the well and everything else on the site. Even after the fire burned out, petroleum flooded from the hole for another six months. Energy firms guessed that the blowout hinted at the presence of big new oil-and-gas deposits. Earlier assessments had missed them because of their great depth. Investors rushed in and began to drill.
To McKelveyan social scientists, such stories demonstrate that oil reserves should not be thought of as physical entities. Rather, they are economic judgments: how much petroleum experts believe can be harvested from given areas at an affordable price. Even as companies drain off the easy oil, innovation keeps pushing down the cost of getting the rest. From this vantage, the race between declining oil and advancing technology determines the size of a reserve—not the number of hydrocarbon molecules in the ground. Companies that scrambled to follow the Kern River gusher found millions of barrels of deep oil, but it was mixed with so much water that they couldn’t stop the wells from flooding. Within a few years, almost all the new rigs ceased operation. The reserve vanished, but the oil remained.
This perspective has a corollary: natural resources cannot be used up. If one deposit gets too expensive to drill, social scientists (most of them economists) say, people will either find cheaper deposits or shift to a different energy source altogether. Because the costliest stuff is left in the ground, there will always be petroleum to mine later. “When will the world’s supply of oil be exhausted?” asked the MIT economist Morris Adelman, perhaps the most important exponent of this view. “The best one-word answer: never.” Effectively, energy supplies are infinite.
Sweeping claims like these make Jean Laherrère’s teeth hurt. Laherrère spent 37 years exploring for oil and gas for the French petroleum company Total before co-founding the Association for the Study of Peak Oil and Gas. ASPO was born after Laherrère and Colin Campbell, another retired petroleum geologist, predicted in 1998 that “within the next decade, the supply of conventional oil will be unable to keep up with demand.” Given the record-high petroleum reserves of the time, the claim was gutsy. Campbell and Laherrère insisted that talk of ever more oil was nonsense. In the 1980s, the Organization of the Petroleum Exporting Countries, the intergovernmental cartel that controls most crude oil, discussed allocating sales on the basis of member states’ reserves: the bigger a nation’s reserves, the more oil OPEC would let that nation sell. In such a system, countries would have every incentive to overstate their holdings. As Campbell and Laherrère noted, six of the 11 OPEC members abruptly hiked their reserve estimates during these discussions. Incredibly, some nations more than doubled their estimates, without a word of explanation for why they now had so much more oil in the ground. (OPEC eventually decided not to allocate oil in this way.) The supposed glut was a charade, Laherrère told me when we spoke in February. The reserves didn’t exist. “We said the [plateau in oil production]would begin before 2010, and we were correct.”
Far from being infinite, Laherrère said, petroleum supplies are finite by definition. The Earth contains only so many hydrocarbon molecules that can be extracted by human effort. “Once we have used up the easy oil, new types of cheap energy will not appear by magic. We will keep drilling for oil, and it will not be easy to get. Look at the enormously expensive equipment they use now only to keep up production.”
Oil prices soared, as if on cue, after Laherrère and Campbell’s prediction. By 2008, they had hit levels unseen since the Carter administration. “The supply of oil is limited,” President George W. Bush declared that year, echoing his predecessor. “There is a growing consensus that the age of cheap oil is coming to an end,” announced the British government’s Energy Research Centre. “A peak of conventional oil production before 2030 appears likely and there is a significant risk of a peak before 2020.” Bookstore shelves shudder beneath the avalanche of warnings: The Big Flatline: Oil and the No-Growth Economy. Peak Oil and the Second Great Depression (2010–2030). The End of Growth. The Crash Course. Peeking at Peak Oil. (All have come out in the past three years.)
McKelveyans remain undeterred. Morris Adelman is in failing health and could not speak to me, but I reached two of his students, Michael Lynch and Philip K. Verleger. Lynch, the president of the energy-consulting firm SEER, agreed with Laherrère that reserve estimates are sometimes manipulated for financial reasons—Shell’s chairman resigned in 2004, after the company was caught misstating its reserves—but didn’t think it mattered much. “Shell is still pumping oil,” he said. “The peak-oil people always say, ‘Look at this super-technological rig—see how expensive the equipment is now.’ I see it and think, Look at how good we’ve gotten at doing this.” Lynch added, “The airlines have jettisoned their wooden biplanes and now use 747s. That’s not because we’re running out of sky and it’s harder to fly. It’s because the technology is getting better and increasing our reach.”
More important, to Verleger’s way of thinking, the peak-oil battle has become irrelevant. Verleger, a former economic official in the Ford and Carter administrations, is now a visiting fellow at the Peterson Institute for International Economics in Washington, D.C. Since Hubbert’s time, the dispute has focused on “conventional” petroleum, the type found in regular oil wells, most of which is in the Middle East and controlled by OPEC. Production of conventional oil has indeed plateaued, as Hubbertians warned: OPEC’s output has remained roughly flat since 2005. In part, the slowdown reflects the diminishing supply of this kind of oil. Another part is due to the global recession, which has stalled demand. But a third factor is that OPEC’s conventional petroleum is being supplemented—and possibly supplanted—by what the industry calls “unconventional” petroleum, which for the moment mainly means oil and natural gas from fracking. Fracking, Verleger says, is creating “the biggest change in energy in almost 100 years—a revolution.” That revolution, in his view, will have a big winner: the United States.
The argument is simple. The need to import expensive foreign oil has been a political and economic burden on the United States for decades. Today, though, fracking is unleashing torrents of oil in North Dakota and Texas—it may create a second boom in the San Joaquin Valley—and floods of natural gas in Pennsylvania, West Virginia, and Ohio. So bright are the fracking prospects that the U.S. may become, if only briefly, the world’s top petroleum producer. (“Saudi America,” crowed The Wall Street Journal. But the parallel is inexact, because the U.S. is likely to consume most of its bonanza at home, rather than exporting it.) Oil may cost more than in the past, but prices will surely stabilize. No more spikes! Still more important, this nation is fracking so much natural gas that its price today is less than a third of its price in Europe and Asia—a big cost advantage for American industry. As companies switch to cheap natural gas, a Citigroup report argued last year, the U.S. petroleum boom could add as much as 3.3 percent to America’s GDP in the next seven years.
Until about 1970, the United States produced almost enough petroleum for its own needs. Then, just as Hubbert predicted, domestic oil production began to wane. Suddenly the United States was vulnerable. OPEC had launched an oil embargo in 1967, but it had next to no effect, because the U.S. produced so much of its own oil. Six years later, with U.S. imports surging, OPEC launched a second embargo. Oil prices quadrupled—and caused a massive panic, complete with fistfights at gas stations that were broadcast and rebroadcast on local TV news. “Energy independence!” was the new call from Washington. Perhaps the only ideal shared by Nixon, Carter, and Reagan, it became the holy grail of American politics. George W. Bush, flanked by Democrats, signed the Energy Independence and Security Act of 2007; Barack Obama, fighting with Republicans, has repeatedly touted the need to “get America closer to energy independence.”
Largely because of little-noticed research by government agencies and small companies, that goal is within sight, says Leonardo Maugeri, a former director of the petrochemical division of the Italian energy firm Eni. The United States will still import oil, he argued last summer in a report from Harvard’s Kennedy School of Government. But domestic production will increase so much that by 2020, all of this country’s oil needs “theoretically could come entirely from the Western Hemisphere.” Within a decade, in other words, the U.S. could, if it wanted, stop importing oil from the Middle East. In November, the International Energy Agency agreed, though it pushed the date of independence to 2035. The fracking-led oil-and-gas boom, Philip Verleger said in January, will lead to an American “economic Renaissance.” The United States will at last escape the world made by Churchill, at least for a while.
Nations like Japan, China, and India will still be stuck in that world, as will much of Europe and Southeast Asia. Many of these nations do not have shale deposits to frack, the requisite technological base, or, even if they have both the shale and the technology, the entrepreneurial infrastructure to finance such sweeping changes. Nonetheless, they want to be freed from their abrasive reliance on OPEC. The United States and Canada, mindful that the good times will not last forever, are also hunting for new supplies. All have been looking with ever-increasing interest at a still-larger energy source: methane hydrate.
The land sheds organic molecules into the water like a ditchdigger taking a shower. Sewage plants, fertilizer-rich farms, dandruffy swimmers—all make their contribution. Plankton and other minute sea beings flourish where the drift is heaviest, at the continental margins. When these creatures die, as all living things must, their bodies drizzle slowly to the seafloor, creating banks of sediment, marine reliquaries that can be many feet deep. Microorganisms feed upon the remains.
In a process familiar to anyone who has seen bubbles coming to the surface of a pond, the microbes emit methane gas as they eat and grow. This undersea methane bubbles up too, but it quickly encounters the extremely cold water in the pores of the sediment. Under the high pressure of these cold depths, water and methane react to each other: water molecules link into crystalline lattices that trap methane molecules. A cubic foot of these lattices can contain as much as 180 cubic feet of methane gas.
Most methane hydrate, including the deposit Japan is examining in the Nankai Trough, is generated in this way. A few high-quality beds accumulate when regular natural gas, the kind made underground by geologic processes, leaks from the earth into the deep ocean. However methane hydrate is created, though, it looks much like everyday ice or snow. It isn’t: ordinary ice cannot be set on fire. More technically, ice crystals are typically hexagonal, whereas methane-hydrate crystals are clusters of 12- or 14-sided structures that in scientists’ diagrams look vaguely like soccer balls. Methane molecules rattle about inside the balls, unable to escape. The crystals don’t dissolve in the sea like ordinary ice, because water pressure and temperature keep them stable at depths below about 1,000 feet. Scientists on the surface refer to them by many names: methane hydrate, of course, but also methane clathrate, gas hydrate, hydromethane, and methane ice.
Estimates of the global supply of methane hydrate range from the equivalent of 100 times more than America’s current annual energy consumption to 3 million times more. A tiny fraction—1 percent or less—is buried in permafrost around the Arctic Circle, mostly in Alaska, Canada, and Siberia. The rest is beneath the waves, a reservoir so huge that some scientists believe sudden releases of undersea methane eons ago set off abrupt, catastrophic changes in climate. Humankind cannot tap into the bulk of these deep, vast deposits by any known means. But even a small proportion of a very big number is a very big number.
Hydrates were regarded purely as laboratory curiosities until the 1930s, when a Texas petroleum researcher realized that they were clogging natural-gas pipelines in cold weather. Three decades later, exploration in Siberia revealed gelid bands of methane hydrate embedded in the tundra. Meanwhile, oceanographers were observing anomalies in sonar readings of the seafloor. Some areas of the bottom bounced sound waves back more sharply than one would expect from muddy sediment. It was like waving a flashlight in a dark room and being startled by the flash from a mirror. Three geologists suggested in 1971 that these reflective zones were layers of methane hydrate. Not until 1982 did researchers obtain a large chunk of methane hydrate—a three-foot section of a core sample. The gas inside was 99.4 percent methane. That year, the United States established a methane-hydrate research program.
The investigation was a small, belated part of a global push into unconventional petroleum that had been spurred by the oil shocks of the 1970s. For civilians, understanding unconventionals is difficult, not least because of the taxonomic hodgepodge the industry uses to describe them: tar sands, tight oil, heavy oil, shale gas, coal-bed methane, shale oil, oil shale. (Exasperatingly, shale oil is different from oil shale.) All of these different flavors of petroleum are “unconventional” simply because in the past they were too hard to pull from the earth to be worth the bother. Nowadays technology has made many of them accessible.
With the odd exception, unconventionals can be broken into two rough categories: forms of petroleum that are heavier and less refined than the crudest of crude oil, and forms that are lighter and more refined than crude oil. Both are worth huge sums and entangled in dispute, much like conventional petroleum. But the second category, which includes the natural gas from methane hydrate, seems likely to play a much larger role in humankind’s future—economically, politically, and, most of all, environmentally.
The first, heavy category consists of petroleum that must be processed on-site to be transformed into oil. Tar sands, for instance, consist of ordinary sand mixed with bitumen, a sludgy black goo that hasn’t withstood enough geological heat and pressure to be converted fully into ordinary oil. The most important tar-sand deposits are underneath an expanse of subarctic forest in central Canada that is roughly the size of England; they make up the third-biggest proven oil reserve in the world. In most cases, mining tar sands involves drilling two horizontal wells, one above the other, into the bitumen layer; injecting massive gouts of high-pressure steam and solvents into the top well, liquefying the bitumen; sucking up the melted bitumen as it drips into the sand around the lower well; and then refining the bitumen into “synthetic crude oil.” Refining in this case includes removing sulfur, which is then stored in million-ton, utterly useless Ozymandian slabs around mines and refineries.
Economists sometimes describe a fuel in terms of its energy return on energy invested (EROEI), a measure of how much energy must be used up to acquire, process, and deliver the fuel in a useful form. OPEC oil, for example, is typically estimated to have an EROEI of 12 to 18, which means that 12 to 18 barrels of oil are produced at the wellhead for every barrel of oil consumed during their production. In this calculation, tar sands look awful: they have an EROEI of 4 to 7. (Steaming out the bitumen also requires a lot of water. Environmentalists ask, with some justification, where it all is going to come from.)
Conveying tar-sands oil to its biggest potential markets, in the United States, will involve building a huge pipeline from Alberta to Texas, which has attracted vituperative opposition from environmental groups and some local governments. The U.S. State Department has long delayed issuing permits to allow this pipeline to cross the border, a stall that has outraged energy boosters, who charge that the Obama administration is spitting in the soup of Canada, America’s most important ally. The boosters say little about the two 100 percent Canadian pipelines—one to shoot tar-sands oil to a port in British Columbia, a second to Montreal—that 100 percent Canadian opposition has stalled. All the while, indigenous groups in central Canada, people armed with special powers granted by the Canadian constitution, have carpet-bombed tar-sands country with lawsuits. Regardless of the merits of the protesters’ arguments, it is hard to believe that they will be completely ineffective, or that tar-sands oil will flow freely anytime soon.
Much more prominent is the second unconventional category, the most important subcategory of which is the natural gas harvested by fracking shale. Every few years, the U.S. government produces a map of American shale beds. Flipping through a time series of these maps is like watching the progress of an epidemic—methane deposits pop up everywhere, and keep spreading. To obtain shale gas, companies first dig wells that reach down thousands of feet. Then, with the absurd agility of anime characters, the drills wriggle sideways to bore thousands of feet more through methane-bearing shale. Once in place, the well injects high-pressure water into the stone, creating hairline cracks. The water is mixed with chemicals and “proppant,” particles of sand or ceramic that help keep the cracks open once they have formed. Gas trapped between layers of shale seeps past the proppant and rises through the well to be collected.
Water-assisted fracturing has been in use since the late 1940s, but it became “fracking” only recently, when it was married with horizontal drilling and the advanced sensing techniques that let it be used deep underground. Energy costs are surprisingly small; a Swiss-American research team calculated in 2011 that the average EROEI for fracked gas in a representative Pennsylvania county was about 87—about six times better than for Persian Gulf oil and 16 times better than for tar sands. (Fracking uses a lot of water, though, and activists charge that the chemicals contaminate underground water supplies.) Because of fracking, U.S. natural-gas reserves have jumped by almost three-quarters since 2000.
Shale gas has its detractors. Far from being a game changer, Jean Laherrère told me, shale gas is a “Ponzi scheme” in which oil companies acquire largely fictional methane deposits to polish their balance sheets for Wall Street. A February study from the Post Carbon Institute, an anti-fossil-fuel think tank, dismissed shale gas as, at best, “a temporary reprieve from having to deal with the real problems”; the group’s general tenor is indicated by the special URL it set up for the report: shalebubble.org. But these views are not widely shared. Two days after I last spoke with Laherrère, the head of the U.S. Energy Information Administration told a congressional hearing that the additions to America’s energy reserves ballyhooed in the agency’s most recent report “were—by a large margin—the highest ever recorded since EIA began publishing proved reserve estimates in 1977.”
As Economics 101 would predict, the arrival of vast quantities of methane from fracking has already made U.S. natural-gas prices plummet. In response, hundreds of wells have shut down, preserving methane deposits that can be tapped someday in the future. But U.S. natural-gas production has hardly been affected. Neither has demand: more and more industries, attracted by low prices, are switching to gas from oil and coal—especially coal.
Today, a fifth of U.S. energy consumption is fueled by coal, mainly from Appalachia and the West, a long-term energy source that has provided jobs for millions, a century-old way of life—and pollution that kills more than 10,000 Americans a year (that estimate is from a 2010 National Research Council study). Roughly speaking, burning coal produces twice as much carbon dioxide as burning the equivalent amount of natural gas. Almost all domestic coal is used to generate electricity—it produces 38 percent of the U.S. power supply. Fracking is swiftly changing this: in 2011, utilities reported plans to shut down 57 of the nation’s 1,287 coal-fired generators the following year. Largely in consequence, U.S. energy-related carbon-dioxide emissions have dropped to figures last seen in 1995. Since 2006, they have fallen more than those from any other nation in the world.
The U.S. coal industry has taken to complaining of a “war on coal.” But the economic hit has been less than one would expect; U.S. coal exports, mainly to Europe, almost doubled from 2009 to 2011. In the sort of development that irresistibly attracts descriptors like ironic, Germany, often touted as an environmental model for its commitment to solar and wind power, has expanded its use of coal, and as a result is steadily increasing its carbon-dioxide output. Unlike Americans, Europeans can’t readily switch to natural gas; Continental nations, which import most of their natural gas, agreed to long-term contracts that tie its price to the price of oil, now quite high. “It’s like someone said, ‘We’ll sell you all the tea you want, based on the price of coffee,’ ” Michael Lynch, the energy consultant, told me. “And you said, ‘What a great idea! I’ll lock myself into it for decades.’ ” He laughed. “Truly, you can’t make this stuff up.”
Here I should confess to personal bias. Twelve years ago, a magazine asked me to write an article about energy supplies. While researching, I met petroleum geologists and engineers who told me about a still-experimental technique called hydraulic fracturing. Intrigued, I asked several prominent energy pundits about it. All scoffed at the notion that it would pay off. To be fair, some early fracking research was outlandish; three early trials involved setting off atomic weapons underground (they did produce natural gas, but it was radioactive). I don’t want to embarrass anyone I spoke with. I failed to exercise independent judgment, and did not mention hydraulic fracturing in my article, so I was just as mistaken. But I also don’t want to miss the boat again. Even though plenty of experts discount methane hydrate, I now am more inclined to pay attention to the geologists and engineers who foresee a second, fracking-type revolution with it, a revolution that—unlike the shale-gas rush, mostly a North American phenomenon—will ripple across the globe.
Japan, which has spent about $700 million on methane-hydrate R&D over the past decade, has the world’s biggest hydrate-research program—or perhaps that should be programs, because provincial governments on Japan’s west coast formed a second hydrate-research consortium last year. (Several researchers told me that the current towel-snapping between Beijing and Tokyo over islands in the East China Sea is due less to nationalistic posturing than to nearby petroleum deposits.) In mid-March, Japan’s Chikyu test ended a week early, after sand got in the well mechanism. But by then the researchers had already retrieved about 4 million cubic feet of natural gas from methane hydrate, at double the expected rate. Japan’s Ministry of Economy, Trade, and Industry is eager to create a domestic oil industry; at present, the nation produces just one one-thousandth of its own needs. Perhaps overoptimistically, the ministry set 2018 as a target date for commercializing methane hydrate. India and South Korea are following along, each spending as much as $30 million a year on hydrate experiments; the Korean program is growing especially aggressively.
By contrast, the U.S. Department of Energy program is small—its annual budget is about $15 million, most of which is devoted to basic research on gas hydrates’ formation and location. About $2.4 million goes to U.S. Geological Survey methane-hydrate researchers, who have been test-mining onshore deposits in frigid Alaska and northwestern Canada. Based in Woods Hole, Massachusetts, and Denver, Colorado, the USGS program has about eight full-time researchers, as well as collaborators from Japan, Canada, Germany, India, and several oil companies.
Although most U.S. research has been in the far north, the most promising U.S. deposits are in the Gulf of Mexico. Hydrates are thought to blanket about 174,000 square miles of the gulf, an area about the size of California. At least part of the deposit, seepage from conventional hydrocarbon reservoirs, is top-quality stuff, though nobody has any idea how much is actually recoverable. What is known, says Timothy Collett, the energy-research director for the USGS program, is that some of the gulf’s more than 3,500 oil and gas wells are in gas-hydrate areas. Extracting these hydrates, in his view, is the logical next step. “To keep feeding the infrastructure, you have to maintain a certain return. Otherwise, you’ll abandon it,” he told me. “For the individual manager of a large installation with a multimillion-dollar budget, it might be well within your interest, as you go into decline on deepwater production, to start looking at gas hydrate.”
If one nation succeeds in producing commercial quantities of undersea methane, others will follow. U.S.-style energy independence, or something like it, may become a reality in much of Asia and West Africa, parts of Europe, most of the Americas. To achieve this dream, history suggests, subsidies to domestic producers will be generous and governments will slap fees on petroleum imports—especially in Asia, where dependence on foreign energy is even more irksome than it is here. In addition to North America, the main sources of conventionally extracted natural gas are Russia, Iran, and Qatar (Saudi Arabia is also an important producer). All will feel the pinch in a methane-hydrate world. If natural gas from methane hydrate becomes plentiful and cheap enough to encourage nations to switch from oil, as the Japanese hope, the risk pool will expand to include Brunei, Iraq, Nigeria, the United Arab Emirates, Venezuela, and other petro-states.
The results in those nations would be turbulent. Petroleum revenues, if they are large, exercise curious and malign effects on their recipients. In 1959, the Netherlands found petroleum on the shores of the North Sea. Money gurgled into the country. To general surprise, the flood of cash led to an economic freeze. Afterward, economists realized that salaries in the new petroleum industry were so high that nobody wanted to work anywhere else. To keep employees, companies in other parts of the economy had to jack up wages, in turn driving up costs. Meanwhile, the surge of foreign money into the Netherlands raised the exchange rate. Soaring costs and currency made it harder for Dutch firms to compete; manufacturing and agriculture faltered; unemployment climbed, except in the oil industry. The windfall led to stagnation—a phenomenon that petroleum cognoscenti now call “Dutch disease.”
Some scholars today doubt how much the Netherlands was actually affected by Dutch disease. Still, the general point is widely accepted. A good modern economy is like a roof with many robust supporting pillars, each a different economic sector. In Dutch-disease scenarios, oil weakens all the pillars but one—the petroleum industry, which bloats steroidally.
Worse, that remaining pillar becomes so big and important that in almost every nation, the government takes it over. (“Almost,” because there is an exception: the United States, the only one of the 62 petroleum-producing nations that allows private entities to control large amounts of oil and gas reserves.) Because the national petroleum company, with its gush of oil revenues, is the center of national economic power, “the ruler typically puts a loyalist in charge,” says Michael Ross, a UCLA political scientist and the author of The Oil Curse (2012). “The possibilities for corruption are endless.” Governments dip into the oil kitty to reward friends and buy off enemies. Sometimes the money goes to simple bribes; in the early 1990s, hundreds of millions of euros from France’s state oil company, Elf Aquitaine, lined the pockets of businessmen and politicians at home and abroad. Often, oil money is funneled into pharaonic development projects: highways and hotels, designer malls and desalination plants. Frequently, it is simply unaccounted for. How much of Venezuela’s oil wealth Hugo Chávez hijacked for his own political purposes is unknown, because his government stopped publishing the relevant income and expenditure figures. Similarly, Ross points out, Saddam Hussein allocated more than half the government’s funds to the Iraq National Oil Company; nobody has any idea what happened to the stash, though, because INOC never released a budget. (Saddam personally directed the nationalization of Iraqi oil in 1972, then leveraged his control of petroleum revenues to seize power from his rivals.)
Shortfalls in oil revenues thus kick away the sole, unsteady support of the state—a cataclysmic event, especially if it happens suddenly. “Think of Saudi Arabia,” says Daron Acemoglu, the MIT economist and a co-author of Why Nations Fail. “How will the royal family contain both the mullahs and the unemployed youth without a slush fund?” And there is nowhere else to turn, because oil has withered all other industry, Dutch-disease-style. Similar questions could be asked of other petro-states in Africa, the Arab world, and central Asia. A methane-hydrate boom could lead to a southwest-to-northeast arc of instability stretching from Venezuela to Nigeria to Saudi Arabia to Kazakhstan to Siberia. It seems fair to say that if autocrats in these places were toppled, most Americans would not mourn. But it seems equally fair to say that they would not necessarily be enthusiastic about their replacements.
Augmenting the instability would be methane hydrate itself, much of which is inconveniently located in areas of disputed sovereignty. “Whenever you find something under the water, you get into struggles over who it belongs to,” says Terry Karl, a Stanford political scientist and the author of the classic The Paradox of Plenty: Oil Booms and Petro-States. Think of the Falkland Islands in the South Atlantic, she says, over which Britain and Argentina went to war 30 years ago and over which they are threatening to fight again. “One of the real reasons that they are such an issue is the belief that either oil or natural gas is offshore.” Methane-hydrate deposits run like crystalline bands through maritime flash points: the Arctic, and waters off West Africa and Southeast Asia.
In a working paper, Michael Ross and a colleague, Erik Voeten of Georgetown University, argue that the regular global flow of petroleum, the biggest commodity in world trade, is also a powerful stabilizing force. Nations dislike depending on international oil, but they play nice and obey the rules because they don’t want to be cut off. By contrast, countries with plenty of energy reserves feel free to throw their weight around. They are “less likely than other states to sign major treaties or join intergovernmental organizations; and they often defy global norms—on human rights, the expropriation of foreign companies, and the financing of foreign terrorism or rebellions.” The implication is sobering: an energy-independent planet would be a world of fractious, autonomous actors, none beholden to the others, with even less cooperation than exists today.
None of this is what makes Christopher Knittel use words like catastrophe. What Knittel is thinking of is, so to speak, the little black specks of Yulin, China. Five years ago, I traveled with a friend to Yulin, in the northwestern province of Shaanxi, not far from Mongolia. We visited the Great Wall, which passes just north of town. In that area, the wall itself had mostly crumbled to nothing, except for the watchtowers, which stuck up every half mile or so. People in one tower were supposed to be able to signal to the next, passing on messages like ships at sea.
When I climbed up one eroded tower, I was surprised to find that I couldn’t see its neighbor. There were little black specks all over my glasses. I cleaned the lenses, but was still unable to make out the next tower. The black specks were not just on my glasses.
Walking around town, my friend and I had noticed that almost every home had a pile of coal outside, soft dark chunks that people shoveled into stoves for cooking and heating. Thousands upon thousands of coal fires were loading the air with tiny dots of soot. Scientists have taken to calling these dots “black carbon,” and have steadily ratcheted up their assessments of its harm. In March, for instance, a research team led by a Mumbai environmental group estimated that black carbon and other particulate matter from India’s coal-fired power plants cause about 100,000 deaths a year.
Environmentalists worry even more about black carbon’s role in climate change. Black carbon in the air absorbs heat and darkens clouds. In some places, it alters rain patterns. Falling on snow, it accelerates melting. A 31-scientist team from nine nations released a comprehensive, four-year assessment in January arguing that planetary black-carbon output is the second-biggest driver of anthropogenic (human-caused) climate change; the little black specks I found on my glasses and clothes have roughly two-thirds the impact of carbon dioxide.
Natural gas produces next to no soot and half the carbon dioxide coal does. In coal-heavy places like China, India, the former Soviet Union, and eastern Europe, heating homes and offices with natural gas instead of coal would be a huge step. An MIT study chaired by Ernest Moniz, whom President Obama nominated for energy secretary in March, called natural gas “a cost-effective bridge” to a “low-carbon future.”
The Chinese government is aware of this, which is one reason it is pursuing both shale gas and methane hydrate. But environmentalists are less enthusiastic than one might imagine about the prospect of weaning ourselves from coal with gas. The reason is that methane itself—unburned natural gas—has a much greater capacity to trap solar heat than carbon dioxide does. (Because methane does not remain in the air as long as carbon dioxide, the precise comparison depends on the chosen time frame; researchers typically say that methane is about 20 or 30 times more potent.) Activists fear that the negative effects of obtaining natural gas could swamp the positive effects of burning it. They are entirely correct, although perhaps not in the way they suppose.
Almost every friend and neighbor I have spoken with about methane hydrate asked whether tapping these undersea deposits could release vast amounts of methane all at once, disastrously altering the planet’s environment. According to Carolyn Ruppel of the Geological Survey, these fears are understandable—but misplaced. If things go awry in a hydrate operation, some of the methane will escape into exactly the cold temperatures and high pressures that trapped it to begin with. Some will be consumed by bacteria, producing carbon dioxide, which dissolves in water; this raises the ocean’s acidity, but not enough to have much effect. Any remaining methane will rise out of the sediment and, like the carbon dioxide, dissolve harmlessly in the ocean. (None of this should be confused with a different source of methane: the decayed vegetation in permafrost, which will release methane if the permafrost thaws.)
The real concern, Ruppel and other researchers told me, is less an explosive methane release from under the Earth’s surface—the environmental disaster that might have caused havoc eons ago—than a slow discharge at ground level, from the machinery that will pull methane hydrate out of the seafloor. The problem already exists with fracking. “The rule of thumb is that if a well leaks more than about 3 percent” of its methane production into the air, “natural gas actually becomes dirtier than coal, from a climate-change perspective,” says Ramez Naam, the author of The Infinite Resource, a just-published book about the race between environmental degradation and technological innovation. “The amazing thing, though, is that we don’t have any data—nobody is required to monitor methane at the well. So there’s just a few studies, which vary tremendously.” Worse still, the aging natural-gas infrastructure is riddled with holes and seeps; early this year, a survey of gas mains along Boston’s 785 miles of road, the first-ever such examination, found 3,356 leaks. Last August, the Environmental Protection Agency amended the Clean Air Act to require well operators to recapture some methane; because nobody knows how much natural gas is gushing into the air, the new rules’ impact is uncertain.
Still, fixing leaks is a task that developed nations can accomplish. “In the United States,” Lynch says, “it is possible to hire inspectors and send them out in white vans to measure methane emissions. They can tell companies to spray more silicone in the wellheads. Maybe the companies will kick and scream about the bureaucracy and cost, but this is something that can be done.”
What we can’t do, or at least not readily, is overcome the laws of economics. More...
By: Janet Pelley March 28, 2013
By 2100, the world population will pass 10 billion people. How can we sustain that population, while throttling back climate change? Some scientists think part of the answer is to plant forests and grow biofuel crops, allowing us to use land to sequester carbon. But those changes in land use also could crowd out food crops, raising the cost of food. In ES&T’s Best Policy Analysis Paper of 2012, researchers model the feedback between the atmosphere, Earth’s ecosystems, and the global economy to show that a fossil fuel tax plus incentives for reforestation and biofuels nearly stabilizes the climate by 2100. Unfortunately, the team reports that the trade-off is a hefty rise in food prices (Environ. Sci. Technol., DOI: 10.1021/es2034729).
John Reilly, an energy economist at Massachusetts Institute of Technology, and his team realized that any analysis of climate policies, such as a carbon tax or incentives for reforestation, is complicated by the fact that land use policy, climate, and the economy are tightly linked. For example, as the climate warms, plant productivity changes, which in turn affects people’s decisions about land use, he says. “Thinking of that complex set of interactions and trying to understand what was going on was the motivation for the study,” Reilly says
Previous studies have narrowly focused on one land use at a time, such as finding the best policies to sequester more carbon in forests but ignoring spillover impacts on cropland. Also these studies did not take into account the interactive and downstream effects of a worldwide carbon tax.“Reducing greenhouse gases with a carbon tax will increase energy costs, and energy is a big input in agricultural production,” Reilly says.
His team decided to study a combination of policies—a carbon tax and incentives for reforestation and biofuels—and ask what the effects would be on the climate, food prices, and land use.
To do so, they developed a set of three models that feed data among one another. First, a global economic model, provided by the MIT team, predicts food prices, land use, and greenhouse gas emissions every five years. Next, those emissions estimates drive a climate model that simulates future climate conditions. These climate outputs, such as carbon dioxide and ozone concentrations, then feed into an ecosystem model, developed by Jerry M. Melillo and his team at the Woods Hole Oceanographic Institution. The ecosystem model generates changes in crop and forest productivity. The economic model then uses these productivity changes, along with expected demand for products produced from the land, to reassess land and energy use, which then determines the next set of greenhouse gas emissions estimates. And the cycle continues.
The team members ran their models through the year 2100 under several different policy scenarios. They found that if humanity sticks with the status quo and attempts no change in climate policy, CO2 concentrations reach 900 ppm by 2100, and global mean temperature hits 5.8 °C above preindustrial levels. However, a worldwide tax on fossil fuel emissions limits the CO2 concentration to 520 ppm and holds temperature rise to 2.7 °C. These metrics are still above the targets set by the 2009 Copenhagen Accord, which are a 2 °C temperature increase and CO2 levels around 450 ppm CO2.
When the team added incentives for reforestation and biofuels to a tax on energy, CO2 concentrations reach only 490 ppm and global temperature increases 2.2 °C over preindustrial levels by the end of the century. But Reilly says these better climate statistics come at a price: Competition for agricultural land from reforestation and biofuels boosts food prices by 80%.
Stephen Polasky, an environmental economist at the University of Minnesota, Twin Cities, says that this paper reveals that policy makers cannot think of energy and food policies as independent.
Even though the scenarios in the paper are far more aggressive than any proposed policies suggested right now, fossil fuel use in the models never drops to levels scientists say we need to stabilize the climate, says Steve Running, a terrestrial carbon scientist at the University of Montana. He also says the paper highlights how any solution to the climate problem requires choices and trade-offs when it comes to land use.
Reilly hopes that others will adopt his team’s modeling practices to find sensible polices to improve the climate.
The authors declare no competing financial interest.
Most efforts to address carbon emissions focus on preventing them from entering the atmosphere in the first place. But how to get rid of CO2 already there? Start-ups are developing prototype air-capture systems.
Introduction
Of all the uncertainties climate change presents, its impact on the production and distribution of food is one of the greatest. We are already feeling the effects: 2012 was a bad year for farmers, with droughts and erratic weather decimating crops and pushing up global food prices. Food prices are at historic highs and there have been two global food crises in the last five years leading to riots in Haiti in 2008 and contributing to the Arab Spring in 2011.
Molly D. Anderson and John Reilly examine the complex challenges and trade-offs humanity faces in a world where climate change is upending traditional assumptions about where and how we can produce enough food for the world’s rapidly growing population.
Molly D. Anderson is a professor at College of the Atlantic and holds the Patrtridge Chair in Food and Sustainable Agriculture Systems.
Until just a few years ago, there were some blithe assumptions about how climate change would affect food security: Like migrating birds, agriculture will simply move north to escape extreme heat, and only food production will be affected by climate change.
Today we recognize that it’s not just temperature, but a whole set of complex interrelated factors — temperature, rainfall, timing, soils, practices throughout the food system and more — that are affected by climate change.
Each crop has its own ideal set of circumstances. Having too many warm nights can be deadly for some crops. Not having enough hours of nighttime freeze can hurt others. Human societies have evolved with agriculture over the last 10,000 years to use particular crops in particular places. Now we’re experimenting with drastic changes in a matter of decades.
It’s not going to be easy, for a number of reasons, for agriculture to just move north. Farming is one of the most place-based occupations in the world. Farmers won’t easily pick up and move north. If they do, the soil they’ll find will be completely different.
Climate change isn’t just affecting the production of food; it’s also affecting consumption of and access to food. Ocean acidification will lower fish catches, which in turn will increase demands on land-based foods. Climate refugees will need new access to food, yet will be unable to produce their own. Food safety will become more challenging.
Food security, as defined by the Food and Agriculture Organization of the UN (FAO), is “when all people, at all times, have physical, social and economic access to sufficient, safe and nutritious food that meets their dietary needs and food preferences for an active and healthy life.”
Even without climate change, with the growing world population, food security will present a challenge. We need to look for win/win solutions — ones that improve food security and sustainability of food systems on the one hand and that mitigate and adapt to climate change on the other.
On the farm level, this means promoting the use of renewable energy in food production, restoring degraded soils and diversifying crops.
On a state and regional level, it means first recognizing food as a basic human right and then making policy decisions that flow from that recognition. For example, using land to produce food would take priority over using land to produce biofuels or animal feed; and states and regions would establish adequate food reserves and be able to set their own food and trade policies. States and regions must promote energy and water efficiency throughout the food system as well.
Globally, we need to slow population growth. One of the most effective ways to do that is to educate girls and women and provide access to contraceptives. We also need to reduce food waste and over-consumption, particularly by wealthy people and nations of the world.
When it comes to food security, the developed world is answering the wrong questions. We’ve focused on increasing availability of food and “feeding the world” (to the benefit of our own corporations). We need to focus on improving food access, reducing our own over-consumption, and addressing why poor people can’t feed themselves in a world with more than enough food for all its inhabitants.
John Reilly is senior lecturer and co-director of the Joint Program on the Science and Policy of Global Change and a senior lecturer at the MIT Sloan School of Management.
Since 1980, the world’s breadbaskets — areas where major crops like maize, wheat, rice, and soy beans are grown — have warmed significantly. Interestingly, the U.S. is the major exception to this global trend. Our agricultural regions have actually experienced somewhat cooler temperatures overall — with a few exceptions.
The effects of climate change on agriculture are likely to be mixed, benefiting crops in some areas and harming crops in others. In colder regions, like New England and much of Canada, growing seasons are becoming longer. We can expect lower crop yields in regions where heat exceeds critical thresholds.
Just as agriculture is a major cause of greenhouse gas emissions, it could also play a major role in mitigating climate change.
Scientists estimate that doubling CO2 concentrations from pre-industrial levels would increase crop yields by as much as 20 to 30 percent, but would also increase the growth of weeds. Furthermore, the increase in crop yields from the effects of more carbon dioxide in the air would largely be offset by the effects of increased temperatures and decreased soil moisture.
Agriculture and climate are both highly complex pieces of the Earth’s ecosystem. Constructing reasonably accurate, useful models of how the two interact is an enormous scientific challenge.
My colleagues at MIT and I have begun developing a model for predicting crop yield changes in the world’s breadbasket regions. We’ve found wide variations in how yields are likely to be affected by climate change. Generally speaking, whether we looked at maize (corn) in North America and West Africa, wheat in Europe and Asia, or soybeans in South America, the results were the same: Areas closer to the equator saw declining yields, some up to 50 percent, while areas closer to the North and South Poles showed increased yields. They balance out at some level, but this kind of change would cause lots of dislocation.
With global population projected to peak at 10 billion sometime after 2050, and with rising incomes allowing more people to eat a resource-intensive diet (i.e. eating more meat), we face great agricultural challenges even without the dislocation and disruption climate change will cause.
Just as agriculture is a major cause of greenhouse gas emissions, it could also play a major role in mitigating climate change. Our studies show that an aggressive global reforestation policy could result in a half-degree Celsius of avoided warming by 2100. The key would be putting a price on carbon for removing carbon dioxide from the atmosphere. This price would create an incentive for landholders to reforest their land, because forests are great absorbers of carbon.
Reforestation comes at a cost. More land for forests means less land for agriculture. That means we could expect to see higher food prices, especially for livestock.
This is part of the trilemma of what to do with land in the 21st century. Do we use it to produce biofuels as a substitute for fossil fuels? Do we use it to produce food? Do we use it to preserve biodiversity and store carbon?
There are unavoidable trade-offs no matter what we decide. There are no easy solutions when it comes to climate change and food security. What is clear is that the worst “solution” would be continued inaction in the face of the overwhelming evidence that climate change has real and growing effects.
On Tuesday, four Democrats in Congress unveiled a brand-new proposal for a carbon tax. The set-up is simple: The U.S. government would slap a fee on fossil-fuel emissions and refund the revenue back to the public.
But there’s a twist: The precise details of the carbon tax have yet to be thrashed out. The four lawmakers are soliciting public comments for how big the tax should be and how best to rebate the money.
The proposal is being put forward by Reps. Henry Waxman and Earl Blumenauer, as well as Sens. Sheldon Whitehouse and Brian Schatz.
Here are the key questions they’re wrestling with:
1. What is the appropriate price per ton for polluters to pay? The draft contains alternative prices of $15, $25 and $35 per ton for discussion purposes.
2. How much should the price per ton increase on an annual basis? The draft contains a range of increases from 2 percent to 8 percent per year for discussion purposes.
3. What are the best ways to return the revenue to the American people? The discussion draft proposes putting the revenue toward the following goals, and solicits comments on how to best accomplish each: (1) mitigating energy costs for consumers, especially low-income consumers; (2) reducing the Federal deficit; (3) protecting jobs of workers at trade-vulnerable, energy intensive industries; (4) reducing the tax liability for individuals and businesses; and (5) investing in other activities to reduce carbon pollution and its effects.
4. How should the carbon fee program interact with state programs that address carbon pollution?
Those are, indeed, difficult questions. So let’s take a look at each of them in turn:
1) How big should the carbon tax be? Economists have long argued that a carbon tax can be an elegant way to tackle climate change. If you tax oil, coal, and natural gas and make them more expensive, then people and companies will either use fewer fossil fuels or seek alternatives. Markets will adjust to the new price.
But there’s plenty of dispute over what the appropriate price on carbon emissions should be. For that, you need to figure out how much damage heat-trapping greenhouse gases are actually causing — and figure out how highly to value future generations. The federal government currently pegs the “social cost of carbon” at $21 per ton. Other economists have concluded that the price should be up to 12 times as much.
2) How quickly does the tax need to rise to curtail emissions? A tax that rises each year should, in theory, drive down emissions. But a lot could depend on how quickly the tax actually rises.
Here’s one example: Sebastian Rausch and John M. Reilly of the MIT Global Change Institute recently put forward a proposal for a $20-per-ton carbon tax that would rise 4 percent each year, starting in 2013. (The funds would be used to offset taxes elsewhere.) Here’s what their model predicts would happen to U.S. greenhouse-gas emissions:
Under this proposal, U.S. greenhouse gas emissions do start declining quite a bit (this is the green line), with a relatively small impact on the U.S. economy. But by 2030, emission levels stall, even though the carbon tax keeps rising by 4 percent each year. The United States wouldn’t get anywhere near the 80 percent cut by 2050 that the White House has envisioned.
It’s possible the MIT model is too pessimistic or wrong. It’s also possible that deeper emissions cuts might require a carbon tax that rises even more sharply. But a higher tax could also prove more costly to the economy unless it’s offset properly. So there’s a delicate trade-off here.
3) What’s the best way to use the carbon tax revenue? A carbon fee usually gets criticized for hurting poorer Americans the most—they spend the biggest slice of their income on gasoline and other energy-intensive products, after all. But Rausch and Reilly found that a lot of the distributional effects depend on how Congress rebates the revenue, as shown in the chart below:
The green line shows how different income groups would be affected in 2015 if the carbon tax was used to fend off cuts to social welfare programs like Medicaid. Lower-income Americans would benefit significantly, while wealthier Americans would take a small hit.
By contrast, the red and blue lines show the effects if revenue from the carbon tax was used to cut the corporate tax or personal income tax—in those cases, higher-income Americans would come out ahead.
If, however, carbon tax revenue was used to cut payroll taxes—that’s the black line—then the welfare effects in 2015 are more or less neutral.
On the flip side, some experts like Mark Muro of Brookings have argued that a portion of the revenue raised by a carbon tax should be used to fund public clean-energy R&D. The country won’t wean itself off oil solely because carbon gets taxed. We’ll also need public-transit alternatives, or electric-vehicle infrastructure, or futuristic new hydrogen cars. And in many cases, Muro argues, the government may have to help bankroll this infrastructure.
4) How does the carbon tax interact with the states? California is currently operating its own comprehensive program to cut greenhouse-gas emissions 80 percent by 2050. And 10 states in the Northeast have a small cap-and-trade program for electric utilities. Should these states somehow get “credit” for moving early on global warming? And what’s the best way to do that under a carbon-tax system?
In any case, these are all difficult questions. Those who want to join in on this debate can submit comments to the lawmakers at cutcarbon@mail.house.gov. The comment period ends April 21.
By Brad Plumer
February 22, 2013
What’s the best way to curtail gasoline consumption? Economists tend to agree on the answer here: Higher gas taxes at the pump are more effective than stricter fuel-economy standards for cars and trucks.
Much more effective, in fact. A new paper from researchers at MIT’s Global Change program finds that higher gas taxes are “at least six to fourteen times” more cost-effective than stricter fuel-economy standards at reducing gasoline consumption.
Why is that? One of the study’s co-authors, Valerie Karplus, offers a basic breakdown here: Fuel-economy standards work slowly, as manufacturers start selling more efficient vehicles, and people retire their older cars and trucks. That turnover takes time. By contrast, a higher gas tax kicks in immediately, giving people incentives to drive less, carpool more, and buy more fuel-efficient vehicles as soon as possible.
A great deal also depends on whether biofuels and other alternative fuels are available. A tax on gasoline makes these alternative fuels more competitive, whereas fuel-economy standards don’t. “We see the steepest jump in economic cost between efficiency standards and the gasoline tax if we assume low-cost biofuels are available,” Karplus said in an MIT press release.
And yet… all this economic research never seems to have any effect on lawmakers. Since 2007, Congress and the Obama administration have moved to increase federal fuel economy standards, now scheduled to rise to 54.5 miles per gallon by 2025. According to the MIT estimates, this will cost the economy six times as much as simply raising the federal gas tax from its current level of 18.4 cents per gallon to 45 cents per gallon. Yet no one in Congress has even proposed the latter option.
One explanation is that the public just prefers things this way. Higher fuel-economy standards do impose costs, but they’re largely “hidden” costs — in the form of pricier vehicles in the showroom. A higher gas tax, by contrast, is visible every time people fill up at the pump.
In fact, a recent NBER paper by MIT’s Christopher Knittel found that this has been the case for decades. Between 1972 and 1980 the price of oil soared 650 percent. There was endless public debate during this period about how best to reduce reliance on fossil fuels. And, as Knittel discovered, the public consistently preferred price controls and fuel-economy standards over higher gas taxes. That was true no matter how often people were informed that gas taxes were the superior option.
“Given the saliency of rationing and vehicle taxes,” Knittel concluded, “it seems difficult to argue that these alternative polices were adopted because they hide their true costs.” In other words, the public seems to have an (expensive) preference for inefficient regulations over higher taxes to curb gasoline. Economists find it maddening, but it’s hard to change.
Further reading:
–On the other hand, if you want to see a rare economic argument for fuel-economy standards, check out this 2006 paper (pdf) by Christopher Knittel. He found that Americans were becoming less sensitive to fuel prices over time — which strengthened the case for policies like CAFE standards.
THE average price of gasoline in the United States, $3.78 on Thursday, has been steadily climbing for more than a month and is approaching the three previous post-recession peaks, in May 2011 and in April and September of last year.
But if our goal is to get Americans to drive less and use more fuel-efficient vehicles, and to reduce air pollution and the emission of greenhouse gases, gas prices need to be even higher. The current federal gasoline tax, 18.4 cents a gallon, has been essentially stable since 1993; in inflation-adjusted terms, it’s fallen by 40 percent since then.
Politicians of both parties understandably fear that raising the gas tax would enrage voters. It certainly wouldn’t make lives easier for struggling families. But the gasoline tax is a tool of energy and transportation policy, not social policy, like the minimum wage.
Instead of penalizing gasoline use, however, the Obama administration chose a familiar and politically easier path: raising fuel-efficiency standards for cars and light trucks. The White House said last year that the gas savings would be comparable to lowering the price of gasoline by $1 a gallon by 2025. But it will have no effect on the 230 million passenger vehicles now on the road.
Greater efficiency packs less of a psychological punch because consumers pay more only when they buy a new car. In contrast, motorists are reminded regularly of the price at the pump. But the new fuel-efficiency standards are far less efficient than raising gasoline prices.
In a paper published online this week in the journal Energy Economics, I and other scientists at the Massachusetts Institute of Technology estimate that the new standards will cost the economy on the whole — for the same reduction in gas use — at least six times more than a federal gas tax of roughly 45 cents per dollar of gasoline. That is because a gas tax provides immediate, direct incentives for drivers to reduce gasoline use, while the efficiency standards must squeeze the reduction out of new vehicles only. The new standards also encourage more driving, not less.
Other industrialized democracies have accepted much higher gas taxes as a price for roads and bridges and now depend on the revenue. In fact, Germany’s gas tax is 18 times higher than the United States’ (and seven times more if the average state gas tax is included). The federal gasoline tax contributed about $25 billion in revenues in 2009.
Raising the tax has generally succeeded only when it was sold as a way to lower the deficit or improve infrastructure or both. A 1-cent federal gasoline tax was created in 1932, during the Depression. In 1983, President Ronald Reagan raised the tax to 9 cents from 4 cents, calling it a “user fee” to finance transportation improvements. The tax rose again, to 14.1 cents in 1990, and to 18.4 cents in 1993, as part of deficit-reduction deals under President George Bush and President Bill Clinton.
A higher gas tax would help fix crumbling highways while also generating money that could help offset the impact on low- and middle-income families. Increasing the tax, as part of a bipartisan budget deal, with a clear explanation to the public of its role in lowering oil imports and improving our air and highways, could be among the most important energy decisions we make.
Valerie J. Karplus is a research scientist in the Joint Program on the Science and Policy of Global Change at M.I.T.
Read more about the study here.
Related: Carbon Tax a 'Win-Win-Win' for America's Future
Nature: Natural hazards: New York vs the sea
By: Jeff Tollefson
February 13, 2013
In the wake of Hurricane Sandy, scientists and officials are trying to protect the largest US city from future floods.
Joe Leader's heart sank as he descended into the South Ferry subway station at the southern tip of Manhattan in New York. It was 8 p.m. on 29 October, and Hurricane Sandy had just made landfall some 150 kilometres south in New Jersey. As chief maintenance officer for the New York city subway system, Leader was out on patrol. He had hoped that the South Ferry station would be a refuge from the storm. Instead, he was greeted by wailing smoke alarms and the roar of gushing water. Three-quarters of the way down the final set of stairs, he pointed his flashlight into the darkness: seawater had already submerged the train platform and was rising a step every minute or two.
“Up until that moment,” Leader recalls, standing on the very same steps, “I thought we were going to be fine.”
Opened in 2009 at a cost of US$545 million, the South Ferry station is now a mess of peeling paint, broken escalators and corroded electrical equipment. Much of Manhattan has returned to normal, but this station, just blocks from one of the world's main financial hubs, could be out of service for 2–3 years. It is just one remnant of a coastal catastrophe wrought by the largest storm in New York's recorded history.
Sandy represents the most significant test yet of the city's claim to be an international leader on the climate front. Working with scientists over the past decade, New York has sought to gird itself against extreme weather and swelling seas and to curb emissions of greenhouse gases — a long-term planning process that few other cities have attempted. But Sandy laid bare the city's vulnerabilities, killing 43 people, leaving thousands homeless, causing an estimated $19 billion in public and private losses and paralysing the financial district. The New York Stock Exchange closed for the first time since 1888, when it was shut down by a massive blizzard.
As the humbled city begins to rebuild, scientists and engineers are trying to assess what happened during Sandy and what problems New York is likely to face in a warmer future. But in a dilemma that echoes wider debates about climate change, there is no consensus about the magnitude of the potential threats — and no agreement about how much the city should spend on coastal defences to reduce them.
On 6 December, during his first major public address after the storm, New York mayor Michael Bloomberg promised to reinvest wisely and to pursue long-term sustainability. But he warned: “We have to live in the real world and make tough decisions based on the costs and benefits.” And he noted that climate change poses threats not just from flooding but also from drought and heat waves. The city must be mindful, he said, “not to fight the last war and miss the new one ahead”.
Calculated risks
In the immediate aftermath of Sandy, lower Manhattan looked like a war zone. Each night, streams of refugees wielding flashlights wandered north out of the blackout zone, where flood waters had knocked out an electrical substation.
The storm devastated several other parts of the city as well. In Staten Island, pounding waves destroyed hundreds of homes, and one neighbourhood in Queens burned to ashes after water sparked an electrical fire. Power outages lasted for more than two weeks in parts of the city. Chastened by the flooding and acutely aware that Hurricane Irene, in 2011, was a near miss, the city is now wondering what comes next.
“Is there a new normal?” asks John Gilbert, chief operating officer of Rudin Management, which manages several office buildings in downtown New York. “And if so, what is it?” Gilbert says that the company is already taking action. At one of its buildings, which took on some 19 million litres of water, the company is moving electrical systems to the second floor. “You have to think that as it has happened, it could happen again,” he says. “And it could be worse.”
At Battery Park, near the South Ferry station, the storm surge from Sandy rose 2.75 metres above the mean high-water level — the highest since gauges were installed there in 1923. In a study published last week in Risk Analysis, researchers working with data from simulated storms concluded that a surge of that magnitude would be expected to hit Battery Park about once every 500 years in the current climate (J. C. J. H. Aerts et al. Risk Anal. http://dx.doi.org/10.1111/risa.12008; 2013).
But the study authors and other scientists say that the real risks may be higher. The study used flooding at Battery Park as a measure of hurricane severity, yet it also showed that some storms could cause less damage there and still hammer the city elsewhere. Factoring in those storms could drive up the probability estimates of major hurricane damage to New York.
The 1-in-500 estimate also does not take into account the unusual nature of Sandy. Dubbed a Frankenstorm, Sandy was a marriage of a tropical cyclone and a powerful winter snowstorm, and it veered into the New Jersey coast along with the high tide of a full Moon. “It was a hybrid storm,” says Kerry Emanuel, a hurricane researcher at the Massachusetts Institute of Technology (MIT) in Cambridge and one of the study's co-authors. “We need to understand how to assess the risks from hybrid events, and I'm not convinced that we do.”
The risks will only increase as the world warms. The New York City Panel on Climate Change's 2010 assessment suggests that local sea level could rise by 0.3–1.4 metres by 2080. Last year, Emanuel and his colleagues found that floods that occur once every 100 years in the current climate could happen every 3–20 years by the end of this century if sea level rises by 1 metre. What is classified as a '500-year' event today could come every 25–240 years (N. Lin et al. Nature Clim. Change 2, 462–467; 2012).
For city planners, the challenge is to rebuild and protect the city in the face of scientific uncertainty. A few scientists have said for more than a decade that the city should armour New York's harbour with a storm-surge barrier similar to the Thames barrier in London. In Sandy's wake, that idea has gained renewed interest, and a New York state panel last month called for a formal assessment of it.
Bridges and barriers
Malcolm Bowman, who heads the storm-surge modelling laboratory at the State University of New York at Stony Brook, has spearheaded the drive for barriers. He imagines a structure roughly 8 kilometres wide and 6 metres high at the entrance to the harbour, and a second barrier where the East River drains into the Long Island Sound. The state panel's cost estimates for such a system range from $7 billion to $29 billion, depending on the design. The harbour barrier could also serve as a bridge for trains and vehicles to the city's airports, suggests Bowman. “My viewpoint is not that we should start pouring concrete next week, but I do think we need to do the studies,” he says. But whether Sandy will push the city to build major defences, Bowman says, “I don't know.”
Disasters have spurred costly action in the past. The 1888 blizzard helped to drive New York to put its elevated commuter trains underground. And in 2012, the US Army Corps of Engineers completed a $1.1-billion surge barrier in New Orleans, Louisiana, as part of a $14.6-billion effort to protect the city after it was battered by hurricanes Katrina and Rita in 2005. But the New York metropolitan area is bigger and more complex than New Orleans, and protecting it will require a multi-pronged approach. Several hundred thousand city residents live along more than 800 kilometres of coastline, and a barrier would not protect much of coastal Long Island, where Sandy wrought considerable damage. Moreover, the barrier would work only against occasional storm surges. It would not hold back the slowly rising sea or protect against flooding caused by rain.
“A storm-surge barrier may be appropriate, but it's never one thing that is going to protect you,” says Adam Freed, a programme director at the Nature Conservancy in New York, who until late last year was deputy director of the city's office of long-term planning and sustainability. “It's going to be a holistic approach, including a lot of unsexy things like elevating electrical equipment out of the basement and providing more back-up generators.”
As part of that holistic effort, officials are exploring options for expanding the remaining bits of wetlands that once surrounded the city and buffered it from storms. In his address, Bloomberg called wetlands “perhaps the best natural barriers against storms that we have”.
But most of the city's wetlands have become prime real estate in recent decades, and Sandy made clear the consequences of developing those areas, says Marit Larson, director of wetlands and riparian restoration for the New York parks department.
A few weeks after the storm, Larson parks her car near the beach on Staten Island and looks out at a field of Phragmites australis, a common marsh reed. The field is part of Staten Island's 'Bluebelt' programme, initiated in the late 1980s to promote wetlands and better manage storm-water runoff. But the patch of wetlands here is smaller than a football pitch, and Sandy's surge rolled over it, damaging the nearby row houses. “If you look at the historical maps,” says Larson, “everything that used to be a wetland got wet.”
New York is now moving to strengthen its network of existing wetlands, which cover some 2,300–4,000 hectares. The mayor's budget plan for 2013–17 includes more than $200 million to restore wetlands as part of an effort to protect and redesign coastal developments.
Sandy also showed how proper construction can help to reduce risks from future storms. In one Staten Island neighbourhood, a battered roof rests on the ground, marking the spot where an ageing bungalow once stood. Next door, a newer house still stands, with no apparent damage apart from a flooded garage — sturdy proof of the value of modern building codes. In New York, newer buildings constructed in 100-year-flood zones, which are defined by the US Federal Emergency Management Agency (FEMA), cannot have any living spaces or major equipment, such as heating units, below the projected flood level (see 'Danger zone').
The city's zoning provisions could not protect against a storm like Sandy: officials estimate that two-thirds of the homes damaged by the storm were outside the 100-year-flood area. But scientists say that the FEMA flood maps were out of date, so even century-scale storms could cause damage well beyond the designated areas. Last month, FEMA began releasing new flood maps for the New York region that substantially expand this zone.
In their latest study, Emanuel and his colleagues estimate the average annual flood risk for New York as only $59 million to $129 million in direct damages. But costs could reach $5 billion for 100-year storms and $11 billion for 500-year storms. These figures do not include lost productivity or damage to major infrastructure, such as subways.
Bowman and other researchers argue that the city should commit to protecting all areas to a 500-year-flood standard, but not all the solutions are physical. A growing chorus of academics and government officials stress that the city must also bolster its response capacity and shore up the basic social services that help people to rebuild and recover.
Most importantly, the city and surrounding region need to develop a comprehensive strategy for defending the coastline, says Jeroen Aerts, a co-author of the Risk Analysis assessment who studies coastal-risk management at VU University in Amsterdam. Aerts is working with New York officials to analyse proposals for the barrier system and a suite of changes in urban planning, zoning and insurance. “You need a master plan,” he says.
“Ultimately, we all have to move together to higher ground.”
Seth Pinsky is working towards that goal. As president of the New York City Economic Development Corporation, he was tapped by Bloomberg to develop a comprehensive recovery plan that will make neighbourhoods and infrastructure safer. He points out that some newer waterfront parks and residential developments along the coast fared well during the storm. For example, at Arverne by the Sea, a housing complex in Queens, Pinsky says that units survived because they are elevated and set back from the water, with some protection from dunes. The buildings suffered little damage compared with surrounding areas.
Intelligent design
The cost of strengthening the city will be astronomical. In January, Congress approved some $60 billion to fund Sandy recovery efforts, with around $33 billion for longer-term investments, including infrastructure repair and construction by the Army Corps of Engineers. Pinsky says that he does not yet know how much of that money will go to New York, but he is sure it will not be enough. The city will define its budget in June, after his group has made its official recommendations. The rebuilding endeavour will probably necessitate a “creative” mix of public and private financing, he says. “It will probably require calling on a combination of almost every tactic that has been tried around the world.”
Even as he calls for more intelligent development, Pinsky says that New York is unlikely to take a drastic approach to dealing with storm surge and sea-level rise. “Retreating from the coastline of New York city both will not be necessary and is not really possible,” he says.
Given the sheer scale of development along the coast, it is hard to argue with Pinsky's assessment. But many climate scientists fear that bolstering coastal developments only delays the eventual reckoning and increases the likelihood of future disasters. The oceans will rise well into the future, they say, so cities will eventually be forced to accommodate the water.
“I don't see anything yet that looks towards long-term solutions,” says Klaus Jacob, a geoscientist at Columbia University's Lamont-Doherty Earth Observatory in Palisades, New York. But Jacob admits that he is as guilty as anyone. In 2003, he and his wife bought a home in a low-lying area on the Hudson River in Piermont, New York. Although it went against his professional principles, he agreed to the purchase with the assumption that he could elevate the house. But height-restriction laws prevented him from doing so, and Sandy flooded the house. The couple are now rebuilding.
“In a way, I think I was in denial about the risk,” Jacob says. He hopes that a new application to raise the house will be approved, but he still fears that the neighbourhood will not survive sea-level rise at the end of the century. New Yorkers and coastal residents everywhere would be wise to learn that lesson. “Ultimately,” Jacob says, “we all have to move together to higher ground.”
Nature 494, 162–164 (14 February 2013) doi:10.1038/494162a
IN THE NEWS: Changing with the climate
MIT News
January 25, 2013
MIT researchers, Massachusetts officials highlight strategies to adapt to climate change.
Just days after President Obama called for action on climate change in his second inaugural address, members of Massachusetts Governor Deval Patrick’s administration joined energy and environment researchers at MIT to discuss strategies for adapting to climate change. The panel discussion on Jan. 23 fostered a continued partnership between MIT and the Commonwealth to advance energy and environment innovation. More....
IN THE NEWS: Reporter's Notebook: An inside tour of the MassDOT
The Tech
January 30, 2013
MIT students frequently use the T and other MassDOT transit systems; since 2010, our IDs even come with a built-in Charlie Card chip. But most students are unfamiliar with the inner workings of the transit system.
Ethan Feuer, Student Activities Coordinator for the MIT Energy Initiative, organized the tour for twenty five students in order to learn more about large infrastructures and emergency preparedness in cities. More....
IN THE NEWS: Climate Research Showcase
MITEI and MIT Joint Program
February 11, 2013
MIT students, researchers help Massachusetts address a post-Sandy world.
MIT students and researchers brought their latest ideas and findings to the table at an event on January 29. The interdisciplinary group of young researchers presented to officials from the Commonwealth’s Executive Office of Energy and the Environment, in hopes that the state would be able to leverage the information for future planning and implementation. More...