Tropical Cyclone Track Forecasting using Fused Deep Learning from Aligned Reanalysis Data

10/23/2019 ∙ by Sophie Giffard-Roisin, et al. ∙ 0

The forecast of tropical cyclone trajectories is crucial for the protection of people and property. Although forecast dynamical models can provide high-precision short-term forecasts, they are computationally demanding, and current statistical forecasting models have much room for improvement given that the database of past hurricanes is constantly growing. Machine learning methods, that can capture non-linearities and complex relations, have only been scarcely tested for this application. We propose a neural network model fusing past trajectory data and reanalysis atmospheric images (wind and pressure 3D fields). We use a moving frame of reference that follows the storm center for the 24h tracking forecast. The network is trained to estimate the longitude and latitude displacement of tropical cyclones and depressions from a large database from both hemispheres (more than 3000 storms since 1979, sampled at a 6 hour frequency). The advantage of the fused network is demonstrated and a comparison with current forecast models shows that deep learning methods could provide a valuable and complementary prediction. Moreover, our method can give a forecast for a new storm in a few seconds, which is an important asset for real-time forecasts compared to traditional forecasts.



There are no comments yet.


page 5

page 18

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Cyclones, hurricanes and typhoons are words designating the same phenomena: a rare and complex event characterized by strong winds surrounding a low pressure area. The ability to forecast their trajectory and intensity forecasts is crucial for the protection of people and property. However, their evolution depends on many factors at different scales, altitudes and times, which leads to modeling difficulties (Emanuel, 2003). As the dynamical models evolve, their forecast accuracy improves; however, historical tropical cyclone databases have scarcely been utilized by machine learning and deep learning methods, to further improve forecast accuracy.

1.1 Existing Storm Forecasts Methods

Today, the forecasts (track and intensity) are provided by numerous guidance models111NHC track and intensity models,, Accessed: 2018-07-04.. Dynamical models solve the physical equations governing motions in the atmosphere and they are influenced by physical models -convective schemes (such as Kain-Fritsch or Simplified Arakawa Schubert), cloud microphysics, land surface model, ocean model, sea/land ice model, planetary boundary layer scheme, surface layer scheme, longwave and shortwave radiation schemes, subgrid-scale diffusion- and by their data assimilation methods (such as 4D-VAR). They are computationally demanding and in current practice older model runs are adjusted in order to be considered early methods, i.e. available in real time. Statistical models, in contrast, are based on historical relationships between storm behavior and various other parameters (DeMaria et al. (2005)). Current forecasts produced by Regional Specialized Meteorological Centers, like the American Official NHC Forecast (OFCL), are driven by consensus or ensemble methods able to combine different dynamical models 11footnotemark: 1 (up to 20 models for the Global Ensemble Forecast System 222GEFS,, Accessed: 2018-12-12.) .

1.2 Deep Learning and Convolutional Neural Networks (CNN)

A convolutional neural network (CNN) is a deep learning architecture widely adopted as a very effective model for analyzing images or image-like data for pattern recognition

(Krizhevsky et al., 2012; Milletari et al., 2016)

. A CNN is structured in layers: an input layer connected to the data, an output layer connected to the quantities to estimate, and multiple hidden layers in between. The hidden layers of a CNN typically consist of convolutional layers, pooling layers, fully connected layers and normalization layers. The convolutional operations are inspired by the cortex visual system, where each neuron only processes data for its receptive field. Fully connected (FC) layers, usually at the end of the network, connect every neuron in one layer to every neuron in another layer. The advantage of CNN is that it can learn to recognize spatial patterns by exploiting translation invariance (i.e. all parts of the image are processed in a similar way), and thus can extract features automatically while considerably reducing the number of parameters.

Recurrent neural networks (RNN) are a class of artificial neural networks that can model temporal dynamic behavior for a time sequence. Unlike feedforward neural networks (like CNNs), RNNs can use their internal state (memory) to process sequences of inputs. An LSTM (long short term memory) network is a particular RNN used in different time-series applications. Even though the long-short-term memory (LSTM) networks are among the most successful methods for predicting time-series events, they are still difficult to train, and simpler CNNs may outperform LSTMs (Bai et al. (2018)). Moreover, encoding time frames as different input channels in a CNN architecture already proved its efficiency if the history size is fixed (de Bezenac et al. (2017)).

1.3 Machine Learning and Deep Learning in Forecasting Problems

Current statistical forecasting models still perform poorly with respect to dynamical models, even though the database made of past tropical cyclones is constantly growing 333NHC track and intensity models,, Accessed: 2018-07-04.. Machine learning methods, which are able to capture non-linearities and complex relations, have only been scarcely tested for tropical cyclone tracking. Yet, they have recently shown their efficiency in a number of various other forecasting tasks. In particular, convolutional neural networks (CNNs) have raised attention as they are suited for large imaging (2D or 3D) data. In (Xingjian et al., 2015), a convolutional LSTM model was used for precipitation forecast. Another recent study predicts the evolution of sea surface temperature maps by combining CNNs with physical knowledge (de Bezenac et al. (2017)). CNNs have also been used for the detection of extreme weather events like tropical cyclones from weather model variables such as integrated water vapor, as in Racah et al. (2017).

Only few preliminary studies have tackled tropical cyclone forecast tracking using machine learning. Two studies used recurrent neural networks from only trajectory information. Moradi Kordmahalleh et al. (2016) was tested on 6h- and 12h-forecast on only 4 tropical cyclones, while Gao et al. (2018) was tested on only Northwest Pacific tracks. Rüttgers et al. (2018) proposes to use a generative adversarial network (GAN) to generate the future atmospheric image (harder problem), but only for a 6 hour prediction. Another study uses storm tracks and reanalysis maps as input for a hybrid CNN - LSTM network in order to learn the (x,y) tracking coordinates (Mudigonda et al. (2017)). While these methods are usually not compared with existing forecasts methods, some of them seem to even perform worse than a baseline of constant speed and direction, see Giffard-Roisin et al. (2018b).

1.4 Frame of Reference

When dealing with image-like data, these studies consider a fixed regional map for tracking storms, of size 160 x 80 deg (longitude/latitude) for Mudigonda et al. (2017) and of the size of the Korean peninsula area (around 30 x 30 deg) for Rüttgers et al. (2018). However, a fixed region for tropical cyclone forecast has three major limitations. First, the tracked storm must stay in the region even though tracks often cross oceans (see Figure 1), forcing the uses of a large region, even if it leads to memory issues (Mudigonda et al. (2017)). Moreover, learning local phenomena on a large and non-centered image can be difficult. Finally, it prevents information transfer between storms coming from different basins or regions, where ground truth data is scarce. In our recent work (Giffard-Roisin et al. (2018b)), we showed the advantage of using a moving reference CNN model for forecasting tropical cyclone tracks 6 hours into the future. This gave roughly a 30km mean error whereas other learning methods gave more than 60km (Moradi Kordmahalleh et al., 2016; Rüttgers et al., 2018) and a constant speed baseline gave 46km mean error. However, a 6h-forecast is of little use for catastrophe planning and it is not possible to compare to existing forecast methods as the smallest standard is 24 hours.

Figure 1: Database: more than 3000 tropical/extra-tropical storm tracks since 1979. Dots = initial position, colors = maximal storm strength according to the Saffir-Simpson scale.

1.5 Contributions

We propose to extend this previous work by using a moving frame of reference that follows the storm center for a 24h-forecast tracking task. We pose the tracking problem as the estimation of the displacement vector,

, between current and future locations. Moreover, we propose to use the reanalysis data as cropped images (25 x 25 degrees) centered on the storm location. That way, the computation time is reduced and we can infer information from storms coming from a large number of tropical cyclone basins from both hemispheres. In particular, our database is made up of more than 3000 storms since 1979, sampled at a 6 hour frequency (more than 90 000 time steps). We include past temporal information by adding the reanalysis maps from previous time steps. We propose a fusion convolutional neural network taking into account past trajectories and different fields from reanalysis images (wind fields and pressure), and we treat each time step of a storm as a training data point. This paper focuses on a 24h-forecast as a proof of concept, and could be easily extended to larger forecast times.

We aim at building an end-to-end model using two types of data (track data and 3D reanalysis) as input. For each time step of each storm, we want to independently estimate its future displacement. After presenting the data, we will show how we designed CNNs to learn from the reanalysis and then improved the result by combining it with history tracks and other 0D features (such as longitude, latitude, and maximal sustained windspeed). Figure 2 summarizes the fusion pipeline that predicts the 24h storm displacement. Lastly, we will show the results on the test set and compare these with current forecast models.

Figure 2: General architecture: the three types of data feed three neural networks trained separately. The final fused network is re-trained before predicting the 24h-forecast displacement.

2 Data Description

2.1 Storm Tracks

The raw storm track data used in this study is composed of more than 3000 tropical and extra-tropical storm tracks since 1979, extracted from the NOAA database IBTrACS (Knapp et al. (2010)), shown in Figure 1. The tracks were produced by multiple governmental agencies, depending on the basin. They are defined by the 6-hourly center locations (latitude and longitude), and the database also includes some associated descriptors such as the windspeed (see Section 2.3). It includes both hemispheres and the number of records per storm varies from 2 to 120 time steps. In total, the database counts more than 90,000 time steps and we used our method to predict the 24-h track forecast for each single time step.

2.2 Reanalysis Data

The trajectory of a storm depends on large-scale atmospheric flows. We chose to extract analyzed atmospheric fields from reanalysis data, not the forecast fields. We used the ERA-Interim database (Dee et al. (2011)), which is one of the reanalysis datasets covering the data-rich period since 1979. Reanalysis is a systematic approach to produce datasets for climate monitoring and research, covering the entire globe from the Earth’s surface to well above the stratosphere and estimate hundreds of available variables. ERA-Interim is a global atmospheric reanalysis produced by the European Centre for Medium-Range Weather Forecasts (ECMWF) and is produced in near to real time. The spectral resolution is T255 (around 80 km), the time resolution is 6 hours, and there are 60 vertical pressure levels until 0.1 hPa (altitude around 64 km).

2.3 Feature Selection

In this work, we used storm track data and reanalysis outputs to forecast tropical cyclone tracks. We can classify them into 4 types of information:

  • Past displacements (1D). We define a displacement as the values between the locations of a storm’s center, as recorded in the storm track data, at different times. The time difference, , being in a multiple of 6 hours. The historical displacements of a storm help predict its future displacement . We used the current displacement (i.e. between times and ) and the past displacement (between and ). These features are 1D in the sense that they are defined for each past time step (1D temporal data).

  • Meta data (0D)

    . We chose the following useful features extracted from the IBTrACS database: the current center-point latitude and longitude, the current windspeed at the center of the storm, the current distance to land, and the Jday predictor (Gaussian function of

    Julian day of storm init - peak day of the tropical cyclone season in the hemisphere, see DeMaria et al. (2005)). We refer to such features as 0D because they are not defined on a spatial grid.

  • Wind fields and (spatial fields, 3D).

    We applied a sparse feature selection technique (Automatic Relevance Determination, based on linear regression to the target displacement shift) over the 10 available reanalysis fields on pressure levels, which highlighted the usefulness of two reanalysis fields in particular: wind fields and the geopotential height. Wind fields are the direct observations of the atmospheric flows, so their importance is clear. In order to have a moving frame of reference, we extracted the wind fields of the neighborhood of the storm at every time step from the ERA-interim reanalysis database, see Figure

    3. Specifically, we extracted the u-wind and v-wind fields on a 25x25 degree grid centered on the current storm location, at three atmospheric pressure levels (700 hPa, 500 hPa, and 225 hPa). The choice of the three pressure levels was inspired by the literature on statistical forecast models (DeMaria et al. (2005)) and on a sensitivity analysis. The size of the grid was motivated by the fact that meaningful information is extractable from the movement of air masses around the storm, while focusing on the storm only (10x10 deg) bounds most historic extreme storms (which was confirmed by preliminary training experiments).

    Figure 3: Global atmospheric grids centered on the storm location: wind fields (u and v) and geopotential height (z).
  • Geopotential height fields (spatial fields, 3D). As previously mentioned, the geopotential height was also found relevant for this task from the ARD regression. Similar to wind fields, we extracted the geopotential height (or iso-pressure latitude) fields of the neighborhood of the storm at every time step on a 25x25 degree grid centered on the current storm location, at three atmospheric pressure levels (700 hPa, 500 hPa, and 225 hPa).

In order to capture the dynamics, we extracted the wind fields and the geopotential height measured at times and at the same location. These fields are thus 3D (spatial) x 1D (temporal). We point out that we first used surface reanalysis data, including sea surface temperature, sea level pressure and 10 meter winds, but because of no significant impacts to the result, we concentrated our efforts to atmospheric wind and geopotential fields.

2.4 Set Separation

The storms were randomly separated in three sets as following: training (60%) / validation (20%) / testing (20%). Thus, the storms in the test set have never been seen before by the learning algorithm. Then, within each set, all time instants were treated independently. The training set was used for optimizing the parameters of the neural networks (back-propagation). The validation set was used to select the architecture of the network (Section 3). Finally, the test set was kept hidden and was only used to show the final prediction accuracy at test-time (Section 4).

3 Methodology: a Deep Fusion Model

3.1 Overview

Because of the differing nature of the data sources, it is not straightforward to mix all the data into a neural network (NN); different learning rates are needed. We propose a new fusion NN architecture taking into account the four sources of information. An overview of the architecture we developed is shown in Figure 2. We divided our fusion architecture into three branches: a Wind CNN, a Pressure CNN and a Past tracks + meta NN. The Wind CNN and Pressure CNN are 2D CNNs that take atmospheric fields (long, lat, stacked over height and time) as input, while the Past tracks + meta NN is a small neural network which takes 0D features as input (stacked over time). Each branch of the network makes its predictions independently. We train the parameters of each individual branch of the network for the same task, i.e. predicting the 24h-forecast track. We then integrate the three networks into a fused network and fine-tune the parameters. The different steps will be outlined in the following sections.

3.2 Convolutional Neural Network for Reanalysis

We propose two similar CNN networks for the wind and the pressure fields. We separate them into two networks because the type of data is different and thus different learning rates were needed. We stacked the data over height (pressure level) and time, such that the inputs of the CNNs consist of multiple 2D (long, lat) frames or channels. The Pressure CNN has six input channels (each one of size 25x25), while the Wind CNN input consists of 12 channels ( and

are stacked). We used a typical CNN architecture, alternating convolutional layers (Conv layer) and max-pooling layers, with fully connected layers at the end

(Simonyan and Zisserman, 2014)

. Following conventional wisdom in the computer vision literature, all hidden layers are equipped with the rectification (ReLU) non-linearity and batch normalization. The different configurations that we have evaluated for Wind CNN and Pressure CNN are outlined in Table

2, one per column. All configurations follow the generic design described above and differ only in depth, which is determined by the number of convolutional layers. As shown in Table 2, in order to have fair comparisons among the architectures, we designed configurations with approximately the same number of parameters to estimate.

ConvNet Configurations
7 layers 8 layers 9 layers 10 layers
input (12 channels of size 25*25)
Table 2: Number of parameters (in millions) of the four network configurations tested in Table 2.
Network A B C D
Number of parameters (x ) 2.27 2.33 2.75 2.67
Table 1: Different configurations of Wind CNN tested. The depth of the configuration increases from left to right, as more layers are added. conv3-32 indicates a convolution of size 3x3 with 32 output features. FC means fully connected layer. maxpool indicates a 3x3 max-pooling layer. The ReLU activation and batch normalization layers (applied after each conv. or FC layer) are not shown in the figure.

We evaluated the performance on 24-hour storm track prediction for the Wind CNN. The result of the architecture evaluation on the validation set is shown in Table 3. We give two scores: Root Mean Square Error (RMSE) and Mean Absolute Error (MAE), in kilometers. With the increase of model depth, there is no clear improvement on the result. Since adding more convolutional layers allows the network to learn features at more levels of abstraction, we chose the intermediate Network C.

Model (from Table. 2) Root Mean Square Error () Mean Absolute Error()
A 177.2 145.4
B 178.2 146.6
C 177.6 145.6
D 178.2 146.7
Table 3: Performance of candidate configurations (Wind CNN) on 24 hours storm track prediction, on the validation set using wind fields.

We also evaluated how adding more historical features from past time steps in the input data can improve performance. In addition to and , we did not observe any noticeable improvement by including more data from the same location at previous time steps. We thus only kept the times and .

3.3 Past tracks + meta Neural Network

Another important source of information are the previous displacements and the other IBTrACS features (see Section 2). They can be treated as a size-9 vector of 0D components. We designed a small NN of two small fully connected layers (the green branch in Figure 2) to learn the future track from the 0D features. We use two past displacements from to and from to because more past tracks did not improve the performance.

3.4 Fused Neural Network for Wind, Altitude, and Tracks

Because of the differing natures between the wind fields, pressure fields and past track data, it is not straightforward to mix them as an input of a NN. Indeed, our preliminary experimentation on training a network combining these three types of inputs simultaneously did not give satisfactory enough results. Instead, we first train separately the three individual branches of the network. We then concatenate their two last layers and add a layer at the end of the network (see Figure 2). The new concatenated layers consist of the same weights as before in each branch, plus new connections from each branch to the other ones, which we initialize to zero. That way, the function computed is (at start) the same as previously. We then re-train the whole fused network by allowing every weight to be re-optimized. The number of fused layers (here two) was determined by comparing four different configurations on the validation set, and a different learning rate was tuned on the validation data set for this final optimization.

3.5 Algorithmic Details

We trained our networks using the root mean square error (RMSE) in kilometers between the forecast and the true storm location at

as the loss function. We added an L2 penalty on the weights of the model (


). The training was performed by the Adam optimizer, and each model converged within 200 epochs. Every evaluation was repeated three times and an average score was computed in order to assess the robustness to the random weights initialization. Although the training takes nearly 8 hours using PyTorch 4.0 on 4 TitanX GPUs with data parallelism (

Krizhevsky (2014)), the testing time or inference is only a few seconds.

4 Experimental Evaluation

4.1 Results on the Whole Dataset (all basins)

We have compared the fused network, fusing all three branches, with the three single branches of the networks. Figure 4 shows the 24h-forecast results on the test set, which was 14,256 time steps in total, in absolute distance error. We can see the improvement of fusing networks (mean error : 130 km) with respect to the Wind CNN (mean: 148.9 km), the Pressure CNN (mean: 172.7 km) and the Past tracks + meta NN (mean: 186.6 km) alone. We can also see the importance of separately pre-training the three networks before the fusion, as it improves the mean result by 5 km.

Figure 4: Comparison between the three simple networks (the 0D Neural Network, the Pressure CNN and the Wind CNN), the fused network without separate pre-training (gray), and the fused network with pre-training (red, proposed method). 24h-forecast results on the test set (storms coming from all oceanic basins), in distance between predicted and real locations.

We have also calculated a persistence forecast baseline: a 24-hour prediction that is four times the storm’s last displacement from to . The mean error of this baseline on the test set is 196 km, which is more than 60 km higher than our method.

Moreover, if we only test on tropical cyclone time steps excluding depressions, which are storms of lower intensity, our mean prediction error drops from 130 km to 109.3 km. Observe in Figure 5(a) the global trend, showing that tracks from more intense storms are predicted with a lower mean error than less intense storms. The mean error from tropical cyclones of categories 4 and 5 is below 90km. Figure 5(b) shows the forecast errors with respect to the current distance to land. We can see that a small distance to land, 200km or less, is one of the factors impacting the prediction quality. Lastly, we can see in Table 4 for the results on the test set for the different regions or basins that the best results are in the North Atlantic with a mean error of 130.2km, or 26.5% of the 24h displacement mean distance. The larger error is found in the South Pacific basin, but it is also the basin where we have the smaller number of samples.

Figure 5: 24h-forecast mean errors on the whole test set with respect to (a) the current Saffir-Simpson hurricane category (a higher category means a stronger hurricane, dep means tropical depression, storm means tropical storm); (b) its current distance to land.
mean error (km) rel. to mean disp. (%) Num. time points
North Atlantic 130.2 26.5% 2413
West Pacific 136.1 27.9% 4080
East Pacific 106.9 29.3% 2142
South Pacific 161.7 41.6% 693
North Indian 138.9 51.3% 2286
South Indian 136.1 41.2% 3050
Table 4: 24h-forecast results for the different regions (basins), on the test set. Mean error in km and relative mean error wrt. the mean 24h displacement distance.

4.2 Comparison with Statistical/Consensus Forecasts Methods

We also compared our fusion model CNN with two existing forecasting models: CLP5444best track decay, combination of CLIPER5 (Climatology and Persistence model 5 day) and Decay-SHIFOR (Decay Statistical Hurricane Intensity Forecast)., a statistical model which is often used to benchmark other storm track forecasting methods, and OFCL, the National Hurricane Center official forecast (consensus of dynamical models)555National Hurricane Center Forecast Verification,, Accessed: 2018-07-31.. We extracted the CLP5 prediction results of years 1989-2016 in the Atlantic and Eastern Pacific basins. We compare in Table 5

our fused network with the statistical CLP5 on the test tropical cyclone time instants at which both methods provided a forecast. This means we compared only when there is a one-to-one correspondence, which is 4349 time steps from 258 storms. On both basins, our fused network performs better than the CLP5 model on average and in standard deviation. Moreover, the frequency of forecast errors larger to 200km, or

busts, is also lower for our method, especially in the Atlantic (10% compared to 18%). Such comparison is not possible with the OFCL as this model is modified every year and they only provide forecasts of the version N of the model for the year N. We don’t know the performance of the recent OFCL models on previous years and it would be unfair for them to compare with old results that were potentially obtained with earlier, less efficient models.

Model Atlantic errors (km) East Pacific errors (km)
mean std busts busts mean std busts busts
CLP5 125 90 18.3% 9.8% 112 78 4.4% 2.3%
Fusion 112 71 10.3% 4.5% 88 52 4.1% 1.9%
Table 5: Mean and standard deviation 24h-forecast errors for the Atlantic and Pacific basins on the subset of the test set where both predictions were available (total = 4349 time steps). Busts correspond to the ratio of track errors exceeding 200km (and 250km).

That is why we compared the yearly results of our fused network performance with the two models on the same subset of the test set. These results (mean and standard deviation) per year are shown in Figure 6. The number of storms and time steps each year for this comparison is presented in Figure 7. From the plot, we can see that our fused network behaves better than the statistical approach CLP5 in most of the years. Our deep learning model performs better than the OFCL forecast until year 2010 for the Pacific basin (2006 for the Atlantic). During the 2010s, the OFCL method improved and its mean errors per year were smaller than ours. We can also notice that none of the large error peaks (ATL: 1993, 2003, 2012; EPAC: 1993, 2009, 2013) involve our model, which seems to indicate that our method is robust.

Figure 6: Yearly average of 24-hours storm track forecasting errors (km) and standard deviation on the test set (top figure for storms in Atlantic, bottom figure for storms in East Pacific) for our fused network forecasts (blue), the CLP5 model forecasts (green) and the official NHC forecasts (red), 1989-2016.
Figure 7: Number of storms and timesteps used to compare in the two basins (Atlantic and East Pacific) for every year, 1989-2016.

Finally, we qualitatively compared the predictions with both OFCL and CLP5 models for recent storms of the test set, such as Tropical Cyclone Odile in 2014 (Figure 8), Tropical Cyclone Hermine in 2016 (Figure 9), and Tropical Cyclone Blas in 2016 (Figure 10). The small bars connect each pair of predicted and ground truth location after 24 hours. The longer the length, the larger the error. Even though the official OFCL model has globally smaller forecast errors, on some time points our model outperforms the OFCL. It seems that our method still perform poorly on the land (see Figure 9). A future improvement could be to add the sea/land map as additional feature. Moreover, the three forecasts often have different directions. A neural network model can thus help the current forecast modellers by providing a complementary prediction that could be integrated in a consensus method, as their mistakes are different.

Figure 8: 24-h forecast errors (4 time steps ahead) on Tropical Cyclone Odile in 2016. The bars connect each pair of predicted and ground truth location. The longer the length, the larger the error. At the beginning, the forecasts were not always available (a complete absence of an error bar should be interpreted as no forecast).
Figure 9: 24-h forecast errors (4 time steps ahead) on Tropical Cyclone Hermine in 2016. The bars connect each pair of predicted and ground truth location. The longer the length, the larger the error. At the beginning and at the end of the track, the forecasts were not always available (a complete absence of an error bar should be interpreted as no forecast).
Figure 10: 24-h forecast errors (4 time steps ahead) on Tropical Cyclone Blas in 2016. The bars connect each pair of predicted and ground truth location. The longer the length, the larger the error. At the beginning and at the end of the track, the forecasts were not always available (a complete absence of an error bar should be interpreted as no forecast).

5 Discussion

Our method only needs to be trained once, although this training can be improved with more data. After that, only a few seconds are needed to give a forecast for a new storm because prediction or inference using such models is much faster than training them. This is a significant time improvement over dynamical models, whose bottleneck is the computing speed. However, one has to keep in mind that our method needs current and past reanalysis fields. While they are usually quickly calculated, within few hours, it does increase the total forecast time accordingly.

We have shown a proof-of-concept for 24-hour forecasting, and Giffard-Roisin et al. (2018b) shows that the 6-hour results are also very satisfactory. Yet, more long-term forecasts could be made using the same structure. We conjecture that for very long forecasts, larger than 25x25 degree images might be needed. Moreover, we worked here on trajectory prediction, yet this model can be easily modified by changing the last layer and be trained for another task, such as intensity prediction (see Giffard-Roisin et al. (2018a)).

Other useful features could be found by using different reanalysis fields. Although our choice of wind and geopotential height fields was driven by an automated feature selection method, we did not test all the possible configurations at every pressure level. Potentially, a more refined selection could increase the overall performance. As an example, for the intensity prediction, we think that surface fields such as sea surface temperature should be reconsidered. We could also represent the wind field by streamfunction and velocity potential as opposed to u- and v-wind components, which might help to have less correlated features. Moreover, while the machine learning algorithm could learn the differences of flow direction between North and South, a future improvement could be to flip the fields North-South and to change the sign of the vwind component. The recent release of the new version of ERA reanalysis, ERA 5, might also increase the accuracy. As Hodges et al. (2017) show, the mean offset in tropical cyclone center position in the ERA-Interm reanalysis product can be up to 1 degree for the period from 1979 to 2012, so moving to ERA 5 and using the GFDL Vortex Tracker (Marchok, 2002) would increase our performance. A comparison to other baseline forecasts, such as TVCN (Track Variable ConseNsus), would also be interesting. Finally, our method could be easily transferred to operational Numerical Weather Prediction data by filtering it to the same spatial resolution.

6 Conclusion

We designed a neural network for the 24h-cyclone storm track forecasting using a moving frame of reference that makes use of a common dataset and a unique trained NN for every tropical cyclone of both hemispheres. When a new tropical cyclone occurs, our network can give a forecast in only few seconds. We demonstrated the benefit of coupling past displacements and aligned reanalysis images. Moreover, we also compared with traditional forecasting methods and showed the improvement with respect to the statistical CLP5 model. This is only a proof-of-concept of deep learning for tropical cyclone forecasting, yet we think that such a different approach as machine learning and NN can be very beneficial if integrated in a consensus method.


We thank the National Oceanic and Atmospheric Administration (NOAA) and the European Centre for Medium-Range Weather Forecasts (ECMWF) for their online databases IBTrACS and ERA-Interim.


  • Bai et al. (2018) Bai, S., Kolter, J. Z., and Koltun, V. (2018). An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271
  • de Bezenac et al. (2017) de Bezenac, M., Pajot, A., and Gallinari, P. (2017). Deep learning for physical processes: Incorporating prior scientific knowledge. arXiv preprint arXiv:1711.07970
  • Dee et al. (2011) Dee, D. P., Uppala, S. M., Simmons, A., Berrisford, P., Poli, P., Kobayashi, S., et al. (2011). The era-interim reanalysis: Configuration and performance of the data assimilation system. Quarterly Journal of the royal meteorological society 137, 553–597
  • DeMaria et al. (2005) DeMaria, M., Mainelli, M., Shay, L. K., Knaff, J. A., and Kaplan, J. (2005). Further improvements to the statistical hurricane intensity prediction scheme (ships). Weather and Forecasting 20, 531–543
  • Emanuel (2003) Emanuel, K. (2003). Tropical cyclones. Annual review of earth and planetary sciences 31, 75–104
  • Gao et al. (2018) Gao, S., Zhao, P., Pan, B., Li, Y., Zhou, M., Xu, J., et al. (2018). A nowcasting model for the prediction of typhoon tracks based on a long short term memory neural network. Acta Oceanologica Sinica 37, 8–12
  • Giffard-Roisin et al. (2018a) Giffard-Roisin, S., Gagne, D., Boucaud, A., Kégl, B., Yang, M., Charpiat, G., et al. (2018a). The 2018 climate informatics hackathon: Hurricane intensity forecast. In 8th International Workshop on Climate Informatics. XVII–XX
  • Giffard-Roisin et al. (2018b) Giffard-Roisin, S., Yang, M., Charpiat, G., Kégl, B., and Monteleoni, C. (2018b). Fused deep learning for hurricane track forecast from reanalysis data. In Climate Informatics Workshop Proceedings 2018 (Boulder, United States)
  • Hodges et al. (2017) Hodges, K., Cobb, A., and Vidale, P. L. (2017). How well are tropical cyclones represented in reanalysis datasets? Journal of Climate 30, 5243–5264
  • Knapp et al. (2010) Knapp, K. R., Kruk, M. C., Levinson, D. H., Diamond, H. J., and Neumann, C. J. (2010). The international best track archive for climate stewardship (ibtracs) unifying tropical cyclone data. Bulletin of the American Meteorological Society 91, 363–376
  • Krizhevsky (2014) Krizhevsky, A. (2014). One weird trick for parallelizing convolutional neural networks. arXiv preprint arXiv:1404.5997
  • Krizhevsky et al. (2012) Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems. 1097–1105
  • Marchok (2002) Marchok, T. P. (2002). How the ncep tropical cyclone tracker works. In Preprints, 25th Conf. on Hurricanes and Tropical Meteorology, San Diego, CA, Amer. Meteor. Soc. P. vol. 1
  • Milletari et al. (2016) Milletari, F., Navab, N., and Ahmadi, S.-A. (2016). V-net: Fully convolutional neural networks for volumetric medical image segmentation. In 2016 Fourth International Conference on 3D Vision (3DV) (IEEE), 565–571
  • Moradi Kordmahalleh et al. (2016) Moradi Kordmahalleh, M., Gorji Sefidmazgi, M., and Homaifar, A. (2016). A sparse recurrent neural network for trajectory prediction of atlantic hurricanes. In

    Proceedings of the Genetic and Evolutionary Computation Conference 2016

    (ACM), 957–964
  • Mudigonda et al. (2017) Mudigonda, M., Kim, S., Mahesh, A., Kahou, S., Kashinath, K., Williams, D., et al. (2017). Segmenting and tracking extreme climate events using neural networks. In Deep Learning for Physical Sciences (DLPS) Workshop, held with NIPS Conference
  • Racah et al. (2017) Racah, E., Beckham, C., Maharaj, T., Kahou, S. E., Prabhat, M., and Pal, C. (2017). Extremeweather: A large-scale climate dataset for semi-supervised detection, localization, and understanding of extreme weather events. In Advances in Neural Information Processing Systems. 3402–3413
  • Rüttgers et al. (2018) Rüttgers, M., Lee, S., and You, D. (2018). Prediction of typhoon tracks using a generative adversarial network with observational and meteorological data. arXiv preprint arXiv:1812.01943
  • Simonyan and Zisserman (2014) Simonyan, K. and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
  • Xingjian et al. (2015) Xingjian, S., Chen, Z., Wang, H., Yeung, D.-Y., Wong, W.-K., and Woo, W.-c. (2015). Convolutional lstm network: A machine learning approach for precipitation nowcasting. In Advances in neural information processing systems. 802–810