基于Transformer模型的洞庭湖区水位快速模拟

Rapid simulation of water levels in the Dongting Lake region based on Transformer model

  • 摘要: 为了快速准确模拟洞庭湖区水位变化过程,在长短期记忆神经网络(LSTM)模型基础上,引入Transformer架构和压缩-激励(SE)模块,分别构建了TRANS-LSTM和TRANS-LSTM-SE深度学习模型。采用2010—2023年301个雨量站、6个水位站、8个水文站的6h降水、水位和流量观测资料,对模型进行训练和对比分析。结果表明:①深度学习模型的训练耗时和模拟耗时,远小于同研究区域的水动力学模型;②TRANS-LSTM-SE模型明显优于TRANS-LSTM和LSTM模型,训练期和验证期的Nash-Sutcliffe效率系数均大于0.996;③Transformer架构的多头注意力与SE模块通道注意力机制可协同优化LSTM模型,有效地描述复杂江湖系统内水位非线性时空依赖问题,为洞庭湖区水位快速模拟预报提供了一条新途径。

     

    Abstract: To achieve a rapid and accurate simulation of water level variations in the Dongting Lake region, two deep learning models, i.e., the TRANS-LSTM and TRANS-LSTM-SE models which incorporate the Transformer architecture and the Squeeze-and-Excitation (SE) module, were established based on the Long Short-Term Memory (LSTM) neural network. The proposed models were trained and compared using 6h precipitation, water level, and flow discharge datasets from 301 rainfall stations, 6 water level stations, and 8 hydrological stations from 2010 to 2023, The results indicate that: ① The training and simulation times of the deep learning models are significantly less than those of hydrodynamic models when apply in the same basin; ② The TRANS-LSTM-SE model markedly outperforms both the TRANS-LSTM and the LSTM model, and the Nash-Sutcliffe Efficiency coefficients exceeds 0.996 during both the training and validation periods; ③ The multi-head attention mechanism of the Transformer architecture combined with the SE channel attention mechanism, could synergistically optimize the LSTM model and effectively describe the nonlinear spatiotemporal dependencies of water levels in the complex river-lake system, which provides a new approach for rapid water level simulation and forecasting in the Dongting Lake region.

     

/

返回文章
返回