This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

MMINR: Multi-frame-to-Multi-frame Inference with Noise Resistance for Precipitation Nowcasting with Radar

Feng Sun   Cong Bai College of Computer Science
and Technology
Zhejiang University of Technology
310023 Hangzhou, China
Email: [email protected], [email protected]
   Yi Song SPiesat Information
Technology Company Limited
100195 Beijing, China
Email: [email protected]
   Jinglin Zhang School of Artificial Intelligence
Hebei University of Technology
300401 Tianjin, China
Abstract

Precipitation nowcasting based on radar echo maps is essential in meteorological research. Recently, Convolutional RNNs based methods dominate this field, but they cannot be solved by parallel computation resulting in longer inference time. FCN based methods adopt a multi-frame-to-single-frame inference (MSI) strategy to avoid this problem. They feedback into the model again to predict the next time step to get multi-frame nowcasting results in the prediction phase, which will lead to the accumulation of prediction errors. In addition, precipitation noise is a crucial factor contributing to high prediction errors because of its unpredictability. To address this problem, we propose a novel Multi-frame-to-Multi-frame Inference (MMI) model with Noise Resistance (NR) named MMINR. It avoids error accumulation and resists precipitation noiseś negative effect in parallel computation. NR contains a Noise Dropout Module (NDM) and a Semantic Restore Module (SRM). NDM deliberately dropout noise simple yet efficient, and SRM supplements semantic information of features to alleviate the problem of semantic information mistakenly lost by NDM. Experimental results demonstrate that MMINR can attain competitive scores compared with other SOTAs. The ablation experiments show that the proposed NDM and SRM can solve the aforementioned problems.

I Introduction

Precipitation forecasting, using past rainfall information to predict future rainfall intensity in specific areas, is one of the basic challenges in meteorological research. It is generally used for flood warnings, enhanced airplane flight safety, and so on. In general, this task can be divided into long-term and short-term forecasts. Our paper focuses on the latter, aiming to forecast the future 0-2 hours of precipitation, i.e., precipitation nowcasting.

Numerical Weather Prediction (NWP) [2] is a common method in weather forecasting, which uses large amounts of weather data and supercomputers to simulate the state of the natural atmospheric system. Therefore, it requires expensive computational resources and much complex weather data. At this point, the radar extrapolation-based method is another excellent alternative. It only needs to use radar echo maps to predict, so from another perspective, the precipitation prediction task can also be seen as a computer vision task. Hence, radar echo maps based precipitation nowcasting task can be regarded as a sequence-to-sequence prediction task, in which deep learning based methods have achieved excellent results in computer vision domain.

Currently, most of deep learning-based methods in precipitation nowcasting can roughly be divided into Convolutional Recurrent Neural Network (ConvRNN) [3] based methods and Fully Convolutional Network (FCN) [4] based methods. ConvRNNs can extract both temporal and spatial features simultaneously, which achieves outstanding performance on prediction tasks. However, it calculates feature maps with the step of time, as shown in the (a) of Fig. LABEL:fig:differentModel. This means that it cannot be solved by parallel computation and needs a longer calculation time. In addition, ConvRNN based methods have a relatively complex unit structure, resulting in difficulty for training and being prone to gradient explosion when the model is deeper [6]. In contrast, many FCN based methods can be solved by parallel computation in the training phase with multi-frame as input and single frame as output. In the prediction phase, the single-frame resulting from the previous time step will feedback into the model as input to predict the following step result, as shown in the (b) of Fig. LABEL:fig:differentModel. Due to this strategy, the prediction errors will be accumulated in the recurrent process, resulting in the performance of FCN based methods worse than ConvRNNs. Furthermore, some unpredictable regions, known as precipitation noise, as shown in Fig. LABEL:fig:precipitationNoise, will increase the accumulation of errors. The interval between two consecutive radar frames is often up to several minutes, and with the extreme complexity of the atmospheric system, the sudden vanish or appearance of these noises is almost unpredictable. But these noises attracts much attention from the model and degrades the model performance. Hence, we expect a model that can satisfy the following properties to address the above issues:

  1. (a)

    The model can be solved by parallel computation.

  2. (b)

    The model has a simple architecture.

  3. (c)

    The model can avoid the accumulation of prediction errors.

  4. (d)

    The model can deal with precipitation noise.

This paper proposes a Multi-frame-to-Multi-frame Inference model with Noise Resistance for precipitation nowcasting based on radar echo maps, named MMINR. Since MMINR does not contain any recursive structure and has a simple architecture, it satisfies requirements (a) and (b). Instead of following the circular or single-frame inference model of previous FCN based models, we use multi-frame input to generate multi-frame output at once by adding the output channels in the last convolution layer to reduce the accumulation of prediction errors. So it satisfies the condition (c). Noise Resistance (NR) contains Noise Dropout Module (NDM) and Semantic Restore Module (SRM). NDM reduces noise propagation to the deeper layers of the network by reducing the number of channels layer by layer. To compensate for the semantic information lost by NDM while losing noise, SRM is proposed to enhance the semantic features fusion in the decoder. So MMINR can meet the requirement (d).

To summarize, the contributions of our work are as follows:

  • A novel multi-frame-to-multi-frame inference model named MMINR is proposed for precipitation nowcasting. It has multi-output channels in the last convolution layer. Compared with other FCN based methods, it significantly reduces the accumulation of prediction errors.

  • A new Noise Resistance (NR) strategy is proposed that contains noise dropout module (NDM) in the encoder and semantic restore module (SRM) in the decoder. NDM discards precipitation noise by reducing the features channels layer by layer, while SRM enhances the mining of semantic information by fusion the previous and current time step semantic information.

  • Experimental results on ablation study and comparison with ConvRNN based SOTAs and FCN based SOTAs show that MMINR attains competitive scores. The proposed NDM and SRM alleviate the impact of precipitation noise and enhance the extraction of semantic information.

II Related Works

Refer to caption

n1n_{1}n2n_{2}feature flowupsampledownsampleStage 1NDMCBAMSRMfeature1feature_{1}288×288×256288\times 288\times 256Stage 2NDMCBAMSRMfeature2feature_{2}144×144×128144\times 144\times 128Stage 3NDMCBAMSRMfeature3feature_{3}72×72×6472\times 72\times 64Stage 4NDMCBAMSRMfeature4feature_{4}36×36×3236\times 36\times 32Stage 5NDMCBAMfeature5feature_{5}18×18×3218\times 18\times 32CCconnectfif_{i}fi+1f^{{}^{\prime}}_{i+1}WeakenMatrixBoostMatrixFusionFeatureConv1Conv_{1}Conv2Conv_{2}Conv3Conv_{3}σ\sigma×\timesElement-wiseMultiplication++Element-wiseSummation

Figure 3: The flowchart of MMINR.

II-A ConvRNN based models for precipitation nowcasting

In precipitation nowcasting, ConvRNN based methods achieve an excellent effect. ConvLSTM [3] is the first ConvRNN based method for precipitation nowcasting. It combines convolutional operation with LSTM to extract spatial and temporal information. PredRNN [5] enhances the spatial information extraction ability of the model through extra spatial memory units. PredRNN++ [6] proposed a Causal LSTM unit and gradient highway unit to alleviate the gradient propagation difficulties. MIM [7] decomposes the spatial-temporal information into stationary and non-stationary features and then uses the corresponding unit to predict the features. Although the ConvRNN based methods are designed for spatiotemporal forecasting problem, it suffers from an unavoidable problem that they cannot be computed in parallel, resulting in more extended calculation and training time.

II-B FCN based models for precipitation nowcasting

FCN based models can be solved by parallel computation but lack the sensitivity to temporal information. Many FCN based methods adopt a recurrent strategy. It inputs multi-frame to the model and outputs single frame as prediction result, then it feeds the result into the model again to predict the next frame result and finally multi-frame prediction results can be got. RainNet [8] follows the model of Unet [9], which adopts multi-frame-to-single-frame inference in the training phase and recurrent strategy in the testing phase to achieve multiple frames prediction as output. SmaAt-UNet [10] aims at lightweight. It reduces the number of convolutions in Unet and uses the attention module to compensate for missing semantic information, drastically reducing the model size. However, SmaAt-UNet only predicts one future frame and does not have the capability of continuous prediction. RDCNN [11] contains a recurrent structure based on CNN that is different from LSTM. It auto generates convolution kernel from input data. But RDCNN still does not avoid the accumulation of prediction errors. Therefore, recurrent prediction leading to prediction error accumulation is an urgent problem to be solved.

III Methodology

III-A Problem definition

From the perspective of computer vision, we define the precipitation nowcasting problem as a sequence-to-sequence problem. In this task, the input data and prediction result are the sequences of radar echo maps. XX and Y^\hat{Y} denote the input data and the prediction result respectively. And Y^\hat{Y} has the same size as XX. X={x1,x2,,xN}X=\{x_{1},x_{2},...,x_{N}\} is a collection of NN radar echo maps xix_{i}, each with the same size. Y^={y1^,y2^,,yM^}\hat{Y}=\{\hat{y_{1}},\hat{y_{2}},...,\hat{y_{M}}\} is a collection of MM prediction frames yi^\hat{y_{i}}. So the prediction process can be defined as the following:

Y^=Γ(X).\hat{Y}=\Gamma(X). (1)

where Γ\Gamma represents the precipitation nowcasting model.

III-B Multi-frame-to-Multi-frame Inference (MMI)

ConvRNN based models and FCN based models contain different cyclic structure in precipitation nowcasting, as shown in Fig. LABEL:fig:differentModel. FCN based model receives multiple frames as input and outputs a single frame as a prediction result. Then it feed the prediction result back into the model. Such process can be defined as:

y^n+1=F(x0,x1,,xn),\hat{y}_{n+1}=F(x_{0},x_{1},...,x_{n}), (2)
y^n+2=F(x1,,xn,y^n+1).\hat{y}_{n+2}=F(x_{1},...,x_{n},\hat{y}_{n+1}). (3)

where xix_{i} and y^j\hat{y}_{j} represent the i-th input frame and the j-th prediction result, respectively. FF means FCN based model. Many FCN based models will predict a single frame in the training phase, parallelize computation, and use the recurrent prediction method mentioned above in the testing phase.

ConvRNN based models receive only one frame as input at each time step and uses memory units to transfer features between time steps, as shown in (a) of Fig. LABEL:fig:differentModel. Both FCNs or ConvRNN based methods have the same characteristic. In the testing phase, they take the output from the previous time step as input of the next step. Due to the output at the previous time step must contain the prediction error, the next time step will accumulate error in the recurrent process.

TABLE I: Quantitative evaluation results.
9 frames to 1 frame (9-to-1)
Method CSI/frame \uparrow HSS/frame \uparrow B-MSE \downarrow B-MAE \downarrow
r1 \geq 0.5 r \geq 2 r \geq 5 r \geq 10 r \geq 0.5 r \geq 2 r \geq 5 r \geq 10
SmaAt-UNet [10] (MSI) 0.8321 0.6083 0.3151 0.1290 0.4306 0.3558 0.2056 0.0915 1.8954 0.5359
MMINR-MSI 0.8247 0.6254 0.3399 0.1371 0.4217 0.3628 0.2183 0.0963 1.8286 0.5031
9 frames to 9 frames (9-to-9)
RainNet [8] (MSI + Recurrent) 0.6308 0.3471 0.1268 0.0362 0.3094 0.2093 0.0881 0.0274 6.3938 1.1728
MMINR-MSI+Recurrent2 0.6235 0.3560 0.1321 0.0424 0.3068 0.2169 0.0922 0.0323 5.4057 1.1401
ConvLSTM [3] 0.6784 0.3763 0.1279 0.0242 0.3442 0.2311 0.0900 0.0192 4.8311 0.9695
PredRNN [5] 0.6790 0.3686 0.1265 0.0254 0.3449 0.2271 0.0890 0.0200 4.9645 0.9738
MIM [7] 0.6845 0.3730 0.1261 0.0248 0.3483 0.2297 0.0882 0.0194 4.8804 0.9611
PredRNN++ [6] 0.6817 0.3693 0.1252 0.0274 0.3463 0.2277 0.0876 0.0212 4.9265 0.9710
PFST-LSTM [12] 0.6858 0.3736 0.1326 0.0260 0.3498 0.2303 0.0926 0.0203 4.7077 0.9453
MMINR 0.6769 0.3778 0.1405 0.0412 0.3439 0.2312 0.0970 0.0310 4.9284 0.9750
MMINR-SRM 0.6721 0.3636 0.1294 0.0364 0.3396 0.2228 0.0891 0.0270 4.8176 0.9769
MMINR-SRM-NDM 0.6686 0.3620 0.1209 0.0253 0.3361 0.2228 0.0824 0.0191 4.9516 0.9937
  • 1

    r represents the intensity of rainfall. The unit is mm/h.

  • 2

    This method is only used for fair comparison with RainNet and is not compared with other SOTAs.

TABLE II: rain rate statistic of the KNMI.
Rain Rate (mm/h) Proportion Rainfall Level
0    \leq   x   \leq 0.5 63.7534 % No
0.5  \leq   x   \leq 2 25.6244 % Ligh
2    \leq   x   \leq 5 08.7806 % Ligh to moderate
5    \leq   x   \leq 10 01.5652 % Moderate
10   \leq   x 00.2764 % Moderate to heavy

As shown in (c) of the Fig. LABEL:fig:differentModel, we propose the Multi-frame-to-Multi-frame Inference (MMI) model to avoid error accumulation. It takes a sequence (i.e., multiple frames) as input and directly outputs a sequence. The total process can be written as:

y^n+1,y^n+2,,y^n+m=ϕ(x0,x1,,xn).\hat{y}_{n+1},\hat{y}_{n+2},...,\hat{y}_{n+m}=\phi(x_{0},x_{1},...,x_{n}). (4)

in which ϕ\phi means MMI. nn and mm represent the number of input and output frames. In general, the last layer in other FCN based model are a single convolution operation with one output channel. But in our method, the last convolution layer has mm output channels.

III-C Noise Resistance

Due to the time interval between two continuous radar frames usually up to several minutes, some rainfall areas vanish, arise, or deform suddenly between frames. These areas are almost unpredictable and should be seen as precipitation noises. Those noises will be transmitted to deeper layers, resulting in the model performance degradation in the classical FCN based model, e.g., RainNet [8]. To address this problem, we design a novel Noise Resistance (NR) strategy, which is composed of NDM in encoder and SRM in decoder.

III-C1 NDM

To dropout the noise in the echo map, five stages of processing are constructed, in which each stage contains a NDM. NDM uses convolutional operations to extract features from the input data. Unlike other typical FCN based models, our approach decreases the number of feature channels with the deeper network. More specifically, the first NDM extracts features from the original input radar maps and generates shallow features with 256 channels. The other NDMs further mine the input features and reduce the number of features channels by 1/2 to drop precipitation noise.

NDM drops part of semantic features along with the noise. To alleviate the information loss problem, we add attention mechanism block Convolutional Block Attention Module (CBAM) [13] after each NDM to further mine the semantic features. The process can be defined as:

fi=CBAM(NDM(X)),i{1}f_{i}=CBAM(NDM(X)),i\in\{1\} (5)
fi=CBAM(NDM(Down(fi1))),i{2,3,4,5}f_{i}=CBAM(NDM(Down(f_{i-1}))),i\in\{2,3,4,5\} (6)

where fif_{i} represent the feature from the i-th stage and DownDown means the downsample.

III-C2 SRM

Despite CBAM enhancing the model’s semantic information mining ability, the semantic features dropped in the previous stage still cannot be transferred to the current stage. We designed a Semantic Restore Module (SRM) to retrieve the lost semantic information to deal with this problem. The structure of SRM is shown on the right in Fig. 3.

SRM receives two input data, features of the current stage, record as fif_{i}, and features of the latter stage after upsampling, fi+1f^{{}^{\prime}}_{i+1}. Due to fif_{i} containing the semantic information lost by fi+1f^{{}^{\prime}}_{i+1}, SRM adaptively captures semantic features from fif_{i} to supplement fi+1f^{{}^{\prime}}_{i+1}. Firstly, they are combined by the connect operation. Then three features are extracted by three different convolutional modules, recorded as boost matrix, fusion feature, and weaken matrix. All values in weaken matrix are between 0 and 1 by a sigmoid function. It is multiplied with the fusion feature to reduce some of the feature values that are considered unwanted. The boost matrix is added with the weakened fusion feature to strengthen the values that need to be enhanced.

TABLE III: rain rate statistic of the KNMI.
Category Methods Traning Testing
ConvRNNs ConvLSTM, PredRNN, Recurrent1 Recurrent
MIM, PredRNN++,
PFST-LSTM
FCNs RainNet [8] MSI2 MSI +
Recurrent
FCNs SmaAt-UNet [10] MSI MSI
FCNs Ours MMI MMI
  • 1

    The cyclic structure that feeds the predicted frame back in again.

  • 2

    Multi-frame-to-Single-frame Inference (MSI).

IV Experiments

IV-A Database

Refer to caption
Figure 4: Prediction comparison with different prediction interval.

Our model is trained and tested in a precipitation nowcasting benchmark database proposed by Koninklijk Nederlands Meteorologisch Instituut (KNMI). This database has 420,000 radar echo maps of Netherlands in 5-min intervals and [10] converts the radar maps to rainfall maps. It contains 5734 frame sequences in the training set, 1557 sequences in the test set. We chose randomly 5000 sequences from the default training set as the training set and the rest of 734 as the validation set. Every sequence includes 18 frames, with the size of 288×\times288 pixels. As shown in Table II, we divide the rainfall maps by four thresholds, 0.5 mm/h, 2 mm/h, 5 mm/h, 10 mm/h. They represent light rain, light to moderate rain, moderate rain, and moderate to heavy rain respectively. We train all methods in our experiments on training set. The results of all methods are obtained on testing set. In order to avoid sudden rainfall intensity that harms the training process, we assign 19 mm/h to any rainfall intensity bigger than 19mm/h. For calculation convenience, we deal data with formula as

y=ln(e,x+1)1.51y=\frac{ln(e,x+1)}{1.5}-1 (7)

xx represents raw precipitation data, yy represents the processed data. This formula keeps the data between -1 and 1 and allocates a suitable interval for high rainfall intensity.

IV-B Evaluation Criteria

For comprehensive evaluation, we adopt two binary metrics and two non-binary metrics. Binary metrics binarize the prediction results and focus on the accuracy of prediction results on different rainfall intensity predictions. Non-binary metrics concentrate on the similarity of the prediction results and GT. Critical Success Index (CSI) and Heidke Skill Score (HSS) [14] are binary metrics. CSI measures the fraction of observed and/or forecast events that are correctly predicted and HSS measures the ratio of the odds of making a hit to the odds of making a false alarm. All binary evaluation metrics except correlation are based on true negative (TN), false positive (FP), false negative (FN), and true positive (TP). For binary rainfall map, we define 0.5mm/h, 2mm/h, 5mm/h, 10mm/h as the specific threshold. Due to the extreme imbalance of the precipitation dataset, [15] proposes Balance-MSE (B-MSE) and Balance-MAE (B-MAE) which assign weights to rainfall areas according to rainfall intensities to mitigate the negative impact of data imbalance.

IV-C Implementation Details

PyTorch implements MMINR with a station equipped with NVIDIA GeForce RTX 2080Ti GPU. We set 0.0001 as the initial learning rate and use Adam optimizer for stochastic gradient descent method. We use a mini-batch of 16 sequences. B-MAE is used as the validation loss function. When the validation loss is no longer decreasing during the training phase, the model with the smallest validation loss is selected as model well trained for prediction.

IV-D Quantitative Comparisons with SOTAs

For a comprehensive comparison between MMINR and other SOTAs, as shown in Tab. III, we classify all methods into four categories based on the training and testing strategies used in the original papers. For fair comparison and to explore the model performance, we also provide variants of our models for comparison with other FCN based methods.

IV-D1 Comparisons with FCN based models

For fair comparison with the representative FCN based model RainNet, we provide a variant version of MMINR that adopts multi-frame-to-single-frame inference in the training phase and recurrent strategy in the testing phase to achieve multiple frames prediction. This is the same with RainNet, and we name it as MMINR-MSI+Recurrent. From Tab. I, we can see that MMINR-MSI+Recurrent outperforms RainNet except in light rain prediction. The reason is that MMINR loses a part of light rain information as light rain has largest proportion in the database as shown in Table II. After adopting the MMI structure, MMINR substantially outperforms RainNet in all metrics. This is because our method directly generates a sequence to avoid the accumulation of prediction errors. Besides, we compare MMINR with SmaAt-UNet at nine frames inferring a single frame mode. The data show that MMINR also outperforms SmaAt-UNet.

Refer to caption

T-30 minT-20 minT-10 minT minT+5 minT+15 minT+25 minT+35 minT+45 min2520151050.5

Precipitation (mm/h)

Input

GT

Ours

ConvLSTM

PredRNN

MIM

PredRNN++

PFST-LSTM

RainNet

Figure 5: Visual comparisons of the proposed MMINR and the state-of-the-art methods.

IV-D2 Comparisons with ConvRNN based models

Tab. I also provides quantitative comparison with other ConvRNNs based methods on KNMI dataset. From this table, we can see that MMINR achieves competitive scores among all experimental results. In particular, our methods achieve the best results when the rainfall intensity 2,5,10\geq 2,5,10. Due to the extreme imbalance of the dataset, light rain (i.e., rainfall intensity 0.5\geq 0.5) has a large proportion in KNMI dataset. Therefore some light rain information is lost due to the dropout of precipitation noise, resulting in a weaker performance of our model than other models for light rain prediction. Although our model without the additional temporal information mining structure, but still achieves a close scores to those ConvRNNs based methods.

IV-D3 Comparisons on different prediction interval with SOTAs

We also provide all metrics curves plot with respect to different prediction interval on all methods except SmaAt-UNet, because SmaAt-UNet only predicts one frame. This comparison is shown in Fig. 4. As can be seen in CSI and HSS plots, the proposed MMINR has better results in the early prediction interval. However, when the prediction time goes longer, the performances of all methods deteriorate. But the prediction performance of MMINR is much better than other ConvRNNs based models and also better than RainNet at rainfall intensities greater than 10 mm/h. In the B-MSE and B-MAE plots, the overall effect of MMINR is consistent with other ConvRNN methods and outperforms RainNet, because MMINR avoids the prediction errors accumulation.

IV-E Visual Comparisons with SOTAs

As shown in Fig. 5, we provide a visual comparison of the proposed MMINR with other SOTAs. Our method is more accurate than other ConvRNN models in predicting the direction and the trend of rainfall movement. Besides, our model has better robustness in predicting moderate to heavy rainfall (rainfall intensity greater than 10 mm/h). Although RainNet consistently predicts heavy rainfall, most predictions are wrong, and the predicted rainfall area has severe deformation, which is mainly due to error accumulation.

IV-F Ablation Study

To investigate the importance of NR, we compare MMINR without NR (denoted by MMINR-NDM-SRM) and with NDM (denoted by MMINR-SRM), as shown in Tab. I. From this table, we can see that MMINR-SRM outperforms MMINR-SRM-NDM in all indicators, especially at rainfall intensity above 10 mm/h, with CSI and HSS improving by 43% (0.03640.02530.0253=0.4387\frac{0.0364-0.0253}{0.0253}=0.4387) and 41%, respectively. This also shows that NDM does reduce the model’s attention to noise and improves the ability to mine semantic information.

In addition, comparing MMINR with MMINR-SRM illustrates the importance of SRM. The data show that all metrics results of MMINR are better than without SRM, and all of them have a substantial improvement, which indicates that SRM does play a complementary semantic role.

Finally, comparing MMINR with MMIR without NR (MMINR-NDM-SRM) shows the importance of the proposed NR strategy, in which MMIMR outperforms MMINR-NDM-SRM in all metrics.

V Conclusion

We propose a novel Multi-frame-to-Multi-frame inference framework for precipitation nowcasting task named MMINR. It can be solved by parallel computation in the training and testing phase and avoid the accumulation of prediction errors. Furthermore, considering the precipitation noise in the radar echo maps, which absorb much attention from the deep learning model, we design a Noise Resistance strategy, which contains two modules, Noise Dropout Module and Semantic Restore Module. It reduces effectively the propagation of precipitation noise to the deep layer of the network and mitigates the problem of semantic information loss when dropout the noise. Comprehensive experimental results demonstrate that our method is efficient. For future work, we plan to add a module of temporal information acquisition to compensate for the weak ability of model to mine temporal features.

References

  • [1] C. Wang and Y. Hong, “Application of spatiotemporal predictive learning in precipitation nowcasting,” in AGU Fall Meeting Abstracts, vol. 2018, 2018, pp. H31H–1988.
  • [2] J. Sun, M. Xue, J. W. Wilson, I. Zawadzki, S. P. Ballard, J. Onvlee-Hooimeyer, P. Joe, D. M. Barker, P.-W. Li, B. Golding et al., “Use of nwp for nowcasting convective precipitation: Recent progress and challenges,” Bulletin of the American Meteorological Society, vol. 95, no. 3, pp. 409–426, 2014.
  • [3] X. Shi, Z. Chen, H. Wang, D.-Y. Yeung, W.-K. Wong, and W.-c. Woo, “Convolutional lstm network: A machine learning approach for precipitation nowcasting,” Advances in neural information processing systems, vol. 28, pp. 802–810, 2015.
  • [4] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
  • [5] Y. Wang, M. Long, J. Wang, Z. Gao, and P. S. Yu, “Predrnn: Recurrent neural networks for predictive learning using spatiotemporal lstms,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017, pp. 879–888.
  • [6] Y. Wang, Z. Gao, M. Long, J. Wang, and S. Y. Philip, “Predrnn++: Towards a resolution of the deep-in-time dilemma in spatiotemporal predictive learning,” in International Conference on Machine Learning.   PMLR, 2018, pp. 5123–5132.
  • [7] Y. Wang, J. Zhang, H. Zhu, M. Long, J. Wang, and P. S. Yu, “Memory in memory: A predictive neural network for learning higher-order non-stationarity from spatiotemporal dynamics,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 9154–9162.
  • [8] G. Ayzel, T. Scheffer, and M. Heistermann, “Rainnet v1. 0: a convolutional neural network for radar-based precipitation nowcasting,” Geoscientific Model Development, vol. 13, no. 6, pp. 2631–2644, 2020.
  • [9] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention.   Springer, 2015, pp. 234–241.
  • [10] K. Trebing, T. Stanczyk, and S. Mehrkanoon, “Smaat-unet: Precipitation nowcasting using a small attention-unet architecture,” Pattern Recognition Letters, vol. 145, pp. 178–186, 2021.
  • [11] E. Shi, Q. Li, D. Gu, and Z. Zhao, “A method of weather radar echo extrapolation based on convolutional neural networks,” in International Conference on Multimedia Modeling.   Springer, 2018, pp. 16–28.
  • [12] C. Luo, X. Li, and Y. Ye, “Pfst-lstm: A spatiotemporal lstm model with pseudoflow prediction for precipitation nowcasting,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 14, pp. 843–857, 2020.
  • [13] S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “Cbam: Convolutional block attention module,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 3–19.
  • [14] R. J. Hogan, C. A. Ferro, I. T. Jolliffe, and D. B. Stephenson, “Equitability revisited: Why the “equitable threat score” is not equitable,” Weather and Forecasting, vol. 25, no. 2, pp. 710–726, 2010.
  • [15] X. Shi, Z. Gao, L. Lausen, H. Wang, D.-Y. Yeung, W.-k. Wong, and W.-c. Woo, “Deep learning for precipitation nowcasting: A benchmark and a new model,” in Advances in neural information processing systems, 2017, pp. 5617–5627.