Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror
AI Google

Google's New Hurricane Model Was Breathtakingly Good This Season (arstechnica.com) 43

An anonymous reader quotes a report from Ars Technica: Although Google DeepMind's Weather Lab only started releasing cyclone track forecasts in June, the company's AI forecasting service performed exceptionally well. By contrast, the Global Forecast System model, operated by the US National Weather Service and is based on traditional physics and runs on powerful supercomputers, performed abysmally. The official data comparing forecast model performance will not be published by the National Hurricane Center for a few months. However, Brian McNoldy, a senior researcher at the University of Miami, has already done some preliminary number crunching.

The results are stunning: A little help in reading the graphic is in order. This chart sums up the track forecast accuracy for all 13 named storms in the Atlantic Basin this season, measuring the mean position error at various hours in the forecast, from 0 to 120 hours (five days). On this chart, the lower a line is, the better a model has performed. The dotted black line shows the average forecast error for official forecasts from the 2022 to 2024 seasons. What jumps out is that the United States' premier global model, the GFS (denoted here as AVNI), is by far the worst-performing model. Meanwhile, at the bottom of the chart, in maroon, is the Google DeepMind model (GDMI), performing the best at nearly all forecast hours.

The difference in errors between the US GFS model and Google's DeepMind is remarkable. At five days, the Google forecast had an error of 165 nautical miles compared to 360 nautical miles for the GFS model, more than twice as bad. This is the kind of error that causes forecasters to completely disregard one model in favor of another. But there's more. Google's model was so good that it regularly beat the official forecast from the National Hurricane Center (OFCL), which is produced by human experts looking at a broad array of model data. The AI-based model also beat highly regarded "consensus models," including the TVCN and HCCA products. For more information on various models and their designations, see here.

This discussion has been archived. No new comments can be posted.

Google's New Hurricane Model Was Breathtakingly Good This Season

Comments Filter:
  • by Anonymous Coward

    WayDumberThanDirt told me Trump was banning hurricanes and increasing tariffs 20% on any countries that reported on them.

    Stop The Science !

  • The fact that the linked graph requires a paragraph-long explanation indicates that it is breathtakingly bad.

    • The second sentence explains it perfectly fine to a layperson, and anyone used to interpreting plots has no real problem even without the supplied explanation. The only points of confusion for someone unfamiliar with the different models are the are the labels.

    • According to the graph the best model was off in predicting by 100km after 48 hours. Not breathtaking.
  • by RossCWilliams ( 5513152 ) on Wednesday November 05, 2025 @12:00AM (#65774236)
    I am not sure how much credence to give to a statistical analysis from one season of a handful of hurricanes.
    • by Anonymous Coward

      Agreed. I came here to say the same thing. Hurricane tracks can be fairly predictable but there are some really wild ones that have a mind of their own. I don't think any of the models could predict those mentioned in the Weather Channel's "Strangest Hurricanes" collection. https://weather.com/storms/hur... [weather.com]

    • by AmiMoJo ( 196126 )

      The site seems to be down, but I would hope any academic evaluating this model would have tested it with historic data as well.

      • by habig ( 12787 )
        I suspect that one can't compare the Google results to past seasons because it was trained on those past seasons. So, this is the first abailable year for a validation dataset. They do mention that this year the traditional models performed particularly poorly, but didn't know why.
  • by 93 Escort Wagon ( 326346 ) on Wednesday November 05, 2025 @12:29AM (#65774256)

    They really should've compared it to the European ECMWF. The US's GFS model has fallen way behind the ECMWF in all sorts of ways, and the US government doesn't seem inclined to provide adequate funding to remedy that.

    Note that this is not intended as an anti-Trump post; this has been a longer-term issue over mutliple administrations, both Democratic and Republican.

    • by evanh ( 627108 )

      Google probably just pillaged from ECMWF

    • I think they're not strictly analogous. For example, ECMWF says "TROPICAL CYCLONE MELISSA COULD NOT BE TRACKED" when you try to lookup the tracks for it.
      They may not do tracks for Atlantic cyclones or something.
    • Note that this is not intended as an anti-Trump post

      To be fair, the Trump administration did cut the budget of NOAA and the NWS. [science.org]

    • That got mentioned:

      This early model comparison does not include the “gold standard” traditional, physics-based model produced by the European Centre for Medium-Range Weather Forecasts. However, the ECMWF model typically does not do better on hurricane track forecasts than the hurricane center or consensus models, which weigh several different model outputs. So it is unlikely to be superior to Google’s DeepMind.

    • They really should've compared it to the European ECMWF. The US's GFS model has fallen way behind the ECMWF in all sorts of ways, and the US government doesn't seem inclined to provide adequate funding to remedy that.

      Note that this is not intended as an anti-Trump post; this has been a longer-term issue over mutliple administrations, both Democratic and Republican.

      Yeah, the US model is well known to be worse than the European models overall, but sometimes better in select scenarios. Typically I hear forecasters mixing predictions from the various models based on circumstance.

      But even ignoring that... I'd expect that a trained model like Googles would dominate a relatively tepid hurricane season. What will be interesting is if it does a better job on an extremely atypical, high impact (landfall) storm system, and can replicate that repeatedly. In general in modeling (

  • by ellbee ( 93668 ) on Wednesday November 05, 2025 @01:28AM (#65774294)

    Anyone bashing the GFS is free to improve it - the code is on github, Fortran required. Replicating the data assimilation network is an exercise for the reader.

  • But a season that mostly consists of fish storms seems like not the most challenging task for a prediction model. Who cares what the margin for error is when we're talking hundreds of miles of open ocean?

    The real test will be when we have another season with storms that actually end up heading towards the east coast.

    • Do container ships carrying your Amazon purchases care?

      • Do container ships carrying your Amazon purchases care?

        It's right there in TFS - this was about the Atlantic hurricane season. You might want to look on a map to see where China is. Just sayin'.

        • "Atlantic hurricane season (June 1 to November 30) affects shipping routes by causing delays, rerouting, and increased costs due to port closures, road and rail damage, and dangerous storm conditions. [...]

          Most affected areas
          U.S. Atlantic and Gulf Coasts: States like Florida, Louisiana, Texas, North and South Carolina, and Georgia are particularly vulnerable."

  • "Google's model was so good that it regularly beat the official forecast from the National Hurricane Center (OFCL), which is produced by human experts looking at a broad array of model data."

    This tells me humans do a decent job. Given humans' performance I doubt if it is worth all that comes with GDMI (which may indirectly be causing hurricanes itself).

  • I think the current predictions seem to be remarkably accurate 5-7 days out. It has been a noticeable and dramatic improvement in the last ~5 years.

    If Google's models turn out to be even more accurate, that is simply fantastic!

  • Instead of just the average track error (the dotted black line), I'd be interested in the error of the average track position. In other words, get the track position at each timestamp for all models, average that, then determine the error.
    • by pz ( 113803 )

      Instead of just the average track error (the dotted black line), I'd be interested in the error of the average track position. In other words, get the track position at each timestamp for all models, average that, then determine the error.

      You're describing the consensus models, and they are better than any individual model, at least thus far.

  • the hurricane model probably wasn't nearly as breathtaking as the hurricanes themselves were.
  • Not quite sure how to interpret this without a strong caveat regarding reliability of data sampling in the current environment. There's a lot that's changed within the last year and I lack the expertise to know whether that affects this comparison or not, whether on the overall ranges or the needed sample size to make a conclusion.
  • AI is, at its core, a sophisticated pattern recognition system. It's really good at digesting tons of inputs and spotting patterns. Hurricane prediction is exactly the kind of thing these AI models *should* be good at, given the right kinds of input data, and enough of it.

Every little picofarad has a nanohenry all its own. -- Don Vonada

Working...