David Kristovich, chief editor of the Journal of Applied Meteorology and Climatology, explains how AMS peer review process, as a somewhat private process, ultimately produces the necessary transparency.

Peer review is an unpublished exchange between authors, editors and reviewers. It is meant to assure quality in a journal. At the same time, Kristovich noted that peer review is not just something readers are trusting, blindly. Rather, the peer review process is meant to lead to a more fundamental transparency—namely, it leads to papers that reveal enough to be reproducible.

“Transparency focuses on the way we tend to approach our science,” Kristovich said during an interview at the AMS Publications Commission meeting in Boston in May. “If someone can repeat all of the steps you took in conducting a study, they should come up with the same answer.”

“The most important part of a paper is to clearly define how you did all the important steps. Why did I choose this method? Why didn’t I do this, instead?”

Transparency also is enhanced by revealing information about potential biases, assumptions, and possible errors. This raises fundamental questions about the limits of information one can include in a paper, to cover every aspect of a research project.

“Studies often take years to complete,” Kristovich pointed out. “Realistically, can you put in every step, everything you were thinking about, every day of the study? The answer is, no you can’t. So a big part of the decision process is, ‘What is relevant to the conclusions I ended up with?’”

The transparency of scientific publishing then depends on peer review to uphold this standard, while recognizing that the process of science itself is inherently opaque to the researchers themselves, while they’re doing their work.

“The difference between scientific research and development of a product, or doing a homework assignment—thinking about my kids—is that you don’t know what the real answer is,” Kristovich said. Science “changes your thinking as you move along, so at each step you’re learning what steps you should be taking.”

You can hear the entire interview here.


peer review week banner

At the AMS Annual Meeting panel on Peer Review last January, journal editors Tony Broccoli, Carolyn Reynolds, Walt Robinson, and Jeff Rosenfeld spoke about how authors and reviewers together make good reviews happen:

Robinson: If you want good reviews, and by good I mean insightful and constructive and that are going to help you make your paper better, the way to do that is to write a really good paper. Make sure your ducks are in a row before you send it in. You should have read over that and edited it multiple times. I’m going to, at some point in my life, write a self-help book in which the single word title is, “Edit!” because it applies to many parts of life. Have your colleagues—not even the co-authors—look at it. Buy the person in the office next door a beer to look over the paper and get their comment. There may be problems with the science–and none of our science is ever perfect–but if it’s a really well constructed, well formulated, well written paper, that will elicit really good reviews.

The flip side of that is, if the paper is indecipherable, you’ll get a review back saying, “I’m trying to figure this out” with a lot of questions, and often it’s major revisions. (We don’t reject that many things out of the box.)

The problem is, the author goes back and finally makes the paper at a standard he or she should have sent in the first time. It goes back to the reviewer, and then the reviewer understands the paper and comes back and starts criticizing the science. Then the author gets angry…”You didn’t bring that up the first time!” Well, that’s because the reviewer couldn’t understand the science the first time. So, if you want good, constructive reviews, write good papers!

Reynolds:  You want to make things as easy as possible for the reviewers. Make the English clear, make the figures clear. Allow them to focus on the really important aspects.

Broccoli: I would add, affirming what Walt said, that the best reviews constructively give the authors ideas for making their papers better. Some reviewers are comfortable taking the role as the gatekeeper and trying to say whether this is good enough to pass muster. But then maybe they aren’t as strong as need be at explaining what needs to be done to make the paper good enough. The best reviews are ones that apply high standards but also try to be constructive. They’re the reviewers I want to go back to.

Rosenfeld: I like Walt’s word, “Edit.” Thinking like an editor when you are a reviewer has a lot to do with empathy. In journals, generally, the group of authors is identical or nearly the same as the group of readers, so empathy is relatively easy. It’s less true in BAMS, but it still applies. You have to think like an editor would, “What is the author trying to do here? What is the author trying to say? Why are they not succeeding? What is it that they need to show me?” If you can put yourself in the shoes of the author—or in the case of BAMS, in the shoes of the reader—then you’re going to be able to write an effective review that we can use to initiate a constructive conversation with the author.

Broccoli: That reminds me: Occasionally we see a reviewer trying to coax the author into writing the paper the reviewer would have written, and that’s not the most effective form of review. It’s good to have diverse approaches to science. I would rather the reviewer try to make the author’s approach to the problem communicated better and more sound than trying to say, “This is the way you should have done it.”


peer review week banner

It’s natural that AMS–an organization deeply involved in peer review–participates in Peer Review Week 2017. This annual reflection on peer review was kicked off today by the International Congress of Peer Review and Scientific Publication in Chicago. If you want to follow the presentations there, check out the videos and live streams.

Since peer review is near and dear to AMS, we’ll be posting this week about peer review, in particular the official international theme, “Transparency in Review.”

To help bring some transparency to peer review, AMS Publications Department presented a panel discussion on the process in January at the 2017 AMS Annual Meeting in Seattle. Tony Broccoli, longtime chief editor of the Journal of Climate, was the moderator; other editors on the panel were Carolyn Reynolds and Yvette Richardson of Monthly Weather Review, Walt Robinson of the Journal of Atmospheric Sciences, and Jeff Rosenfeld of the Bulletin of the American Meteorological Society.

You can hear the whole thing online, but we’ll cover parts of the discussion here over the course of the week.

For starters, a lot of authors and readers wonder where editors get peer reviewers for AMS journal papers. The panel offered these answers (slightly edited here because, you know, that’s what editors do):

Richardson: We try to evaluate what different types of expertise are needed to evaluate a paper. That’s probably the first thing. For example, if there’s any kind of data assimilation, then I need a data assimilation expert. If the data assimilation is geared toward severe storms, then I probably need a severe storms expert too. First I try to figure that out.

Sometimes the work is really related to something someone else did, and that person might be a good person to ask. Sometimes looking through what papers they are citing can be a good place to look for reviewers.

And then I try to keep reaching out to different people and keep going after others when they turn me down…. Actually, people are generally very good about agreeing to do reviews and we really have to thank them. It would all be impossible without that.

Reynolds: If you suggest reviewers when you submit to us, I’ll certainly consider them. I usually won’t pick just from the reviewers suggested by the authors. I try to go outside that group as well.

Broccoli: I would add, sometimes if there’s a paper on a topic where there are different points of view, or the topic is yet to be resolved, it can be useful to identify as at least one of the reviewers someone who you know may have a different perspective on that topic. That doesn’t mean you’re going to believe the opinion of that reviewer above the opinion of others but it can be a good way of getting a perspective on a topic.

Rosenfeld: Multidisciplinary papers can present problems for finding the right reviewers. For these papers, I do a lot of literature searching and hunt for that key person who happens to somehow intersect, or be in between disciplines or perspectives; or someone who is a generalist in some way, whose opinion I trust. It’s a tricky process and it’s a double whammy for people who do that kind of research because it’s hard to get a good evaluation.

•  •   •

If you’re interested in becoming a reviewer, the first step is to let AMS know. For more information read the web page here, or submit this form to AMS editors.



Disaster Do-Overs

September 6, 2017 · 0 comments


Ready to do it all over again? Fresh on the heels of a $100+ billion hurricane, we very well may be headed for another soon.

As Houston and the Gulf Coast begin a long recovery from Hurricane Harvey, Hurricane Irma is now rampaging through the Atlantic. With 185 m.p.h. sustained winds on Tuesday, Irma became the strongest hurricane in Atlantic history outside of the Caribbean and Gulf. The hurricane made its first landfall early Wednesday in Barbuda and still threatens the Virgin Islands, Puerto Rico, Cuba, and the United States.


If Irma continues along the general path of 1960’s Hurricane Donna, it could easily tally $50 billion in damage. This estimate, from a study by Karen Clark and Co. (discussed recently on Category 6 Blog), is already four years old (i.e., too low). Increased building costs—which the report notes rise “much faster” than inflation–and continued development could drive recovery costs even higher.


In short, as bad as Houston is suffering, there are do-overs on the horizon—a magnitude of repeated damage costs unthinkable not long ago, before Katrina ($160 million) and Sandy ($70 million).

Repeated megadisasters yield lessons, some of them specific to locale and circumstances. In Miami after Hurricane Andrew, the focus was on building codes as well as the variability of the winds within the storms. After Hurricane Rita, the focus was on improving policies on evacuation. After Hurricane Katrina, while the emergency management community reevaluated its response, the weather community took stock of the whole warnings process. It was frustrating to see that, even with good forecasts, more than a thousand people lost their lives. How could observations and models improve? How could the message be clarified?

Ten years after Katrina, the 2016 AMS Annual Meeting in New Orleans convened a symposium on the lessons of that storm and of the more recent Hurricane Sandy (2012). A number of experts weighed in on progress since 2005. It was clear that challenges remained. Shuyi Chen of the University of Miami, for example, highlighted the need for forecasts of the impacts of weather, not just of the weather itself. She urged the community to base those impacts forecasts on model-produced quantitative uncertainty estimates. She also noted the need for observations to initialize and check models that predict storm surge, which in turn feeds applications for coastal and emergency managers and planners. She noted that such efforts must expand beyond typical meteorological time horizons, incorporating sea level rise and other changes due to climate change.

These life-saving measures are part accomplished and part underway—the sign of a vigorous science enterprise. Weather forecasters continue to hone their craft with so many do-overs. Some mistakes recur. As NOAA social scientist Vankita Brown told the AMS audience about warnings messages at the 2016 Katrina symposium, “Consistency was a problem; not everyone was on the same page.”  Katrina presented a classic problem where the intensity of the storm, as measured in the oft-communicated Saffir-Simpson rating, was not the key to catastrophe in New Orleans. Mentioning categories can actually create confusion. And again, in Hurricane Harvey this was part of the problem with conveying the threat of the rainfall, not just the wind or storm surge. Communications expert Gina Eosco noted that talk about Harvey being “downgraded” after landfall drowned out the critical message about floods.

Hurricane Harvey poses lessons that are more fundamental than the warnings process itself and are eerily reminiscent of the Hurricane Katrina experience: There’s the state of coastal wetlands, of infrastructure; of community resilience before emergency help can arrive. Houston, like New Orleans before it, will be considering development practices, concentrations of vulnerable populations, and more. There are no quick fixes.

In short, as AMS Associate Executive Director William Hooke observes, both storms challenge us to meet the same basic requirement:

The lessons of Houston are no different from the lessons of New Orleans. As a nation, we have to give priority to putting Houston and Houstonians, and others, extending from Corpus Christi to Beaumont and Port Arthur, back on their feet. We can’t afford to rebuild just as before. We have to rebuild better.

All of these challenges, simple or complex, stem from an underlying issue that the Weather Channel’s Bryan Norcross emphatically delineated when evaluating the Katrina experience back in 2007 at an AMS Annual Meeting in San Antonio:

This is the bottom line, and I think all of us in this business should think about this:  The distance between the National Hurricane Center’s understanding of what’s going to happen in a given community and the general public’s is bigger than ever. What happens every time we have a hurricane—every time–is most people are surprised by what happens. Anybody who’s been through this knows that. People in New Orleans were surprised [by Katrina], people in Miami were surprised by Wilma, people [in Texas] were surprised by Rita, and every one of these storms; but the National Hurricane Center is very rarely surprised. They envision what will happen and indeed something very close to that happens. But when that message gets from their minds to the people’s brains at home, there is a disconnect and that disconnect is increasing. It’s not getting less.

Solve that, and facing the next hurricane, and the next, will get a little easier. The challenge is the same every time, and it is, to a great extent, ours. As Norcross pointed out, “If the public is confused, it’s not their fault.”

Hurricanes Harvey and Katrina caused catastrophic floods for different reasons. Ten years from now we may gather as a weather community and enumerate unique lessons of Harvey’s incredible deluge of rain. But the bottom line will be a common challenge: In Hurricane Harvey, like Katrina, a city’s–indeed, a nation’s–entire way of dealing with the inevitable was exposed. Both New Orleans and Houston were disasters waiting to happen, and neither predicament was a secret.

Meteorologists are constantly getting do-overs, like Irma. Sooner or later, Houston will get one, too.



Hurricanes like rapidly-changing Harvey are still full of surprises for forecasters.

The remnants of Caribbean Tropical Storm Harvey made a startling burst Thursday from a tropical depression with 35 mph winds to an 85 mph hurricane in a little more than 12 hours. It has been moving steadily toward a collision with the middle Texas coast and landfall is later Friday. If intensification continues at the same rate, Harvey is likely to be a major hurricane by then, according to a Thursday afternoon advisory from the National Hurricane Center, with sustained winds of 120-125 mph and even higher gusts.

That’s a big “if.”

The drop in central pressure, which had been precipitous all day—a sign of rapid strengthening—had largely slowed by Thursday afternoon. Harvey’s wind speed jumped 50 mph in fits during the same time, but leveled off by late afternoon at about 85 mph. Harvey was a strong Category 1 hurricane on the Saffir-Simpson Hurricane Wind Scale by dinner time.

The intensifying process then slowed. But it turns out this was temporary.

Many signs pointed to continued rapid intensification: a favorable, low-shear environment; expanding upper-air outflow; and warm sea surface temperatures. Overnight and Friday morning, Harvey continued to traverse an eddy of water with high oceanic heat content that has detached from the warm Gulf of Mexico loop current and drifted westward toward the Texas coast. Its impact is apparent as the pressure resumed its plunge and winds have responded, blowing Friday morning at a steady 110 mph with higher gusts.

Further intensification is possible.

In fact, the SHIPS (Statistical Hurricane Intensity Prediction Scheme) Rapid Intensification indices “are incredibly high,” Hurricane Specialist Robbie Berg wrote in the Thursday morning forecast discussion. Guidance from the model then showed a 70 percent chance of another 50 mph jump in wind speed prior to landfall. The afternoon guidance lowered those odds a bit, but still showed a a 64 percent probability.of a 35 mph increase.

It wouldn’t the first time a hurricane has intensified rapidly so close to the Texas coast. In 1999 Hurricane Bret did it, ramping up to Category 4 intensity with 140 mph winds before crashing into sparsely populated Kennedy County and the extreme northern part of Padre Island.

Hurricane Alicia exploded into a major hurricane just prior to lashing Houston in 1983. And 2007’s Hurricane Humberto crashed ashore losing its warm water energy source and capping its intensity at 90 mph just 19 hours after being designated a tropical depression that morning off the northern Texas Coast, a similar boost in intensity as Hurricane Harvey.

Rapid intensification so close to landfall is a hurricane forecasting nightmare. An abundance of peer-reviewed papers reveal that there’s a lot more we need to learn about tropical cyclone intensity, with more than 20 papers published in AMS journals this year alone. Ongoing research into rapidly intensifying storms like Harvey, is helping solve the scientific puzzle, including recent cases such as Typhoon Megi  and Hurricane Patricia. Nonetheless, despite strides in predicting storm motion in past decades, intensification forecasting remains largely an educated guessing game.


Tropical cyclones are usually associated with bad news, but a long-term study about these storms now posted online for publication in BAMS has some good news–about the forecasts, at least. The authors, a Japanese group lead by Munehiko Yamaguchi, studied operational global numerical model forecasts of tropical cyclones since 1991.

Their finding: the model forecasts of storm positions have improved in the last 25 years. In the Western North Pacific, for example, lead time is two-and-a-half days better. Across the globe, errors in 1 – 5 day forecasts dipped by 6 to 14.5 km, depending on the basin.

Here are the improvements for the globe as a whole. Each line is a different modeling center:


While position forecasts with a single model are getting better (not so much with intensity forecasts), it seems natural that the use of a consensus of the best models could improve results even more. But Yamaguchi et al. say that’s not true in every ocean basin. The result is not enhanced in the Southern Indian Ocean, for example. The authors explain:

This would be due to the fact that the difference of the position errors of the best three NWP centers is large rather than comparable with each other and thus limits the impact of a consensus approach.

The authors point towards ways to improve tropical cyclone track forecasts, because not all storms behave the same:

while the mean error is decreasing, there still exist many cases in which the errors are extremely large. In other words, there is still a potential to further reduce the annual average TC position errors by reducing the number of such large-error cases.

For example, take 5-day track forecasts for Hurricane Joaquin in 2015. Hard to find a useful consensus here (the black line is the eventual track):


Yamaguchi et al. note that these are types of situations that warrant more study, and might yield the next leaps in improvement. They note that the range of forecast possibilities now mapped as a “cone of uncertainty” could be improved by adapting to specific situations:

For straight tracks, 90% of the cyclonic disturbances would lie within the cone, but only 39% of the recurving or looping tracks would be within the cone. Thus a situation-dependent track forecast confidence display would be clearly more appropriate.

Check out the article for more of Yamaguchi et al.’s global perspective on how tropical cyclone forecasts have improved, and could continue to improve.


When water-laden air lifts up the eastern slope of the Rockies, enormous thunderstorms and catastrophic flooding can develop. Americans may remember well the sudden, deadly inundation of Boulder, Colorado, in September 2013. For Canadians, however, the big flood that year was in Alberta.

Four years ago this week, 19-23 June 2013, a channel of moist air jetted westward up the Rockies and dumped a foot of rain on parts of Alberta, Canada. The rains eventually spread from thunderstorms along the slopes to a broader stratiform shield. Five people died and 100,000 fled their homes, many in Calgary. At more than $5 billion damage, it was the costliest natural disaster in Canadian history until last year’s Fort McMurray fire.

While we might call it a Canadian disaster, the flood had equally American origins. A new paper in early on-line release for the Journal of Hydrometeorology shows why.

The authors—Yangping Li of the University of Saskatchewan and a team of others from Canadian institutions—focused mostly on how well such mountain storms can be simulated in forecast modeling. But they also traced the origins of the rain water. Local snowmelt and evaporation played a “minor role,” they found. “Overall, the recycling of evaporated water from the U.S. Great Plains and Midwest was the primary source of moisture.”

Here is what the distribution of sources looked like. The colors show net moisture uptake from 6 hours to 7 days before the storm:


Some of the water came from as far east as the Great Lakes, and more than half from the United States. While storms along the eastern slopes of the Rockies often get Gulf of Mexico moisture, in this case, Gulf air had already dumped its moisture on the U.S. Plains. In other words, the soaked Plains merely recycled Gulf moisture back into the air to be carried into Canada.

American floods, Canadian floods, and any combination thereof—Li et al. remind us of the cross-border interdependence of weather, water, and climate … a relationship not just for this week but for the future:

The conditions of surface water availability (e.g. droughts) or agricultural activities over the US Great Plains could exert indirect but potentially significant effects on the development of flood-producing rainfall events over southern Alberta. Future land use changes over the US Great Plains together with climate change could potentially influence these extreme events over the Canadian Prairies.

For more perspectives on this noteworthy flood, take a look at another new paper in early online release–Milrad et al. in Monthly Weather Review–or at the companion papers to the Journal of Hydrometeorology paper: Kochtubajda et al. (2016) and Liu et al. (2016) in Hydrological Processes.


Oceans are deep, and they are integral to the climate system. But the exchanges between ocean and atmosphere that preoccupy many scientists are not in the depths but instead in the shallowest of shallow layers.

A lot happens in the topmost millimeter of the ocean, a film of liquid called the “sea-surface microlayer that is, in many ways, a distinct realm. At this scale, exchanges with the atmosphere are more about diffusion, conduction, and viscosity than turbulence. But the layer is small and difficult to observe undisturbed and over sufficient areas. As a result, “it has been widely ignored in the past,” according to a new paper by Mariana Ribas-Ribas and colleagues in the Journal of Atmospheric and Oceanic Technology.

Nonetheless, Ribas-Ribas and her team, based in Germany, looked for a new way to skim across and sample the critical top 100 micrometers (one tenth of a millimeter) of the ocean. This surface microlayer (SML) “plays a central role in a range of global biogeochemical and climate-related processes.” However, Ribas-Ribas et al. add,

The SML often has remained in a distinct research niche, primarily because it was thought that it did not exist in typical oceanic conditions; furthermore, it is challenging to collect representative SML samples under natural conditions.

In their paper (now in early online release), the authors report on their solution to observing is a newly outfitted remote-controlled catamaran. A set of rotating glass discs with holes scoops up water samples. Pictured below are the catamaran and (at left, top) the glass discs mounted between the hulls and (bottom left) the flow-through system.


Catamarans are not new to this research, but they were generally towed behind other vessels and subject to wake effects or were specialized. The new Sea Surface Scanner (S3) takes advantage of better remote control and power supply technology and can pack multiple sampling and sensors and controls onto one platform. Tests in the Baltic Sea last year showed the ability of S3 to track responses of organisms in the surface microlayer to ocean fronts, upwelling areas, and rainfall. The biological processes in turn affect critical geochemical processes like exchanges of gases and production of aerosols for the atmosphere.

The technology may be a fresh start for research looking in depth at the shallowest of layers. See the journal article for more details on the S3 and its performance in field tests.



by Keith Seitter, AMS Executive Director

President Trump’s speech announcing the U.S. withdrawal from the Paris Climate Agreement emphasizes his assessment of the domestic economic risks of making commitments to climate action. In doing so the President plainly ignores so many other components of the risk calculus that went into the treaty in the first place.

There are, of course, political risks, such as damaging our nation’s diplomatic prestige and relinquishing the benefits of leadership in global economic, environmental, or security matters. But from a scientific viewpoint, it is particularly troubling that the President’s claims cast aside the extensively studied domestic and global economic, health, and ecological risks of inaction on climate change.

President Trump put it quite bluntly: “We will see if we can make a deal that’s fair. And if we can, that’s great. And if we can’t, that’s fine.”

The science emphatically tells us that it is not fine if we can’t. The American Meteorological Society Statement on Climate Change warns that it is “imperative that society respond to a changing climate.” National policies are not enough — the Statement clearly endorses international action to ensure adaptation to, and mitigation of, the ongoing, predominately human-caused change in climate.

In his speech, the President made a clear promise “… to be the cleanest and most environmentally friendly country on Earth … to have the cleanest air … to have the cleanest water.” AMS members have worked long and hard to enable such conditions both in our country and throughout the world. We are ready to provide the scientific expertise the nation will need to realize these goals. AMS members are equally ready to provide the scientific foundation for this nation to thrive as a leader in renewable energy technology and production, as well as to prepare for, respond to, and recover from nature’s most dangerous storms, floods, droughts, and other hazards.

Environmental aspirations, however, that call on some essential scientific capabilities but ignore others are inevitably misguided. AMS members have been instrumental in producing the sound body of scientific evidence that helps characterize the risks of unchecked climate change. The range of possibilities for future climate—built upon study after study—led the AMS Statement to conclude, “Prudence dictates extreme care in accounting for our relationship with the only planet known to be capable of sustaining human life.”

This is the science-based risk calculus upon which our nation’s climate change policy should be based. It is a far more realistic, informative, and actionable perspective than the narrow accounting the President provided in the Rose Garden. It is the science that the President abandoned in his deeply troubling decision.


2017 Atlantic hurricane season namesThe 2017 Atlantic hurricane season has begun. If they haven’t already, people who could be affected by tropical storms and hurricanes should prepare now for the six-month season, which ends November 30 and encompasses the Atlantic Ocean, Caribbean Sea, and Gulf of Mexico.

Here’s what you need to know. Follow the links for more detailed information:

What’s New

Two new products this year will be issued by the National Hurricane Center in addition to the typical tropical storm and hurricane watches, warnings, and advisories:

  • Storm surge watches & warningsStorm surge watches will be issued within 48 hours of possible life-threatening coastal inundation. Storm surge warnings will be issued at least 36 hours before the danger of life-threatening coastal inundation is realized. The new graphical products will be issued in tandem with NHC’s Potential Storm Surge Flooding Map, which quantifies the expected inundation from storm surge and indicates the depth of the flooding on land.
  • Potential tropical storms and hurricanes growing out of disturbances that have not yet become tropical depressions or storms but that pose the threat of bringing tropical storm or hurricane conditions to land areas in 48 hours will be treated as regular storms, with NHC issuing watches, warnings, advisories, and related graphical products as needed.
  • Arrival of tropical storm windsAdditionally, NHC will introduce a map to provide guidance on when users should have their preparations completed before a storm. These experimental graphics will show Time of Arrival of Tropical-Storm-Force Winds—a critical planning threshold for coastal communities. Many preparations become difficult or dangerous once tropical storm conditions begin.


People living in states bordering the Atlantic Ocean and Gulf of Mexico, as well as in The Bahamas, Bermuda, and Caribbean islands, should know that water could be a life-threatening hazard when a hurricane hits. This threat includes storm surge—the sudden rise of seawater at the coast near and to the right of where a hurricane’s center makes landfall. In the United States, hurricane evacuation maps account for the flooding from storm surge and show areas along the coast that people should evacuate if a hurricane threatens. Or it could be from freshwater flooding triggered by a storm’s torrential rain. If either is a risk, make a plan now to leave when threatened.

If water is not a risk, and your area will only face the threat of strong, damaging winds, the official recommendation from the National Hurricane Center is to shelter in place. Meaning, stay put—rather than evacuate. But stay only if your residence is sufficient to weather the storm. Now is the time to prepare your home and property for the possibility of high winds and heavy rain.


A wide variety of academics, private forecasting companies, and the U.S. government issue seasonal hurricane predictions. They generally tend to agree, particularly when a season is expected to be well above or below average. But this year could be a bit different with some predicting fewer tropical storms and hurricanes than is typical in an average year, and others predicting more than the usual numbers of storms and hurricanes. An average Atlantic hurricane season sees a dozen named storms, with a half-dozen of these becoming hurricanes and two going on to become major hurricanes with sustained winds greater than 110 mph.

A quick look at the five longest-running seasonal forecasts for named storms, hurricanes, and major hurricanes, respectively:

The reason for the low numbers in those forecasts with ranges: there’s uncertainty about whether another El Niño may form midway though the season—El Niño is an anomalous warming of the eastern tropical Pacific that causes air to rise and spread out there, creating wind shear over the Atlantic that inhibits tropical development—as well as whether warmer-than-usual sea surface temperatures in the tropical Atlantic favorable for development will remain that way or cool through the season. Increasing numbers in several forecasts updated since earlier in the spring seem to indicate a lower likelihood of both, although forecasters caution they could reduce the numbers if more than a weak El Niño develops and Atlantic SSTs consequently cool.

Bottom Line

Nearly every weather entity states what Acting FEMA Administrator Robert J. Fenton, Jr. recently expressed after NOAA released its hurricane season forecast last week: “Regardless of how many storms develop this year, it only takes one to disrupt our lives.”

The advice of hurricane forecasters and emergency managers alike is to prepare early and as if this will be the year your neighborhood is hit.

Storm Names

If this year’s list of names the National Hurricane Center will use to keep track of Atlantic storms seems familiar, it’s because it is. NHC made it all the way through this list in the 2005 season. The names of five memorable and deadly hurricanes that year were retired: Dennis, Katrina, Rita, Stan, and Wilma. The updated list was used again in 2011, since NHC rotates through six sets of names, and Irene that year also was retired and replaced.