DualMamba: a patch-based model with dual mamba for long-term time series forecasting
Guang-Yu WEI , Hui-Chuan HUANG , Zhi-Qing ZHONG , Wen-Long SUN , Yong-Hao WAN , Ai-Min FENG
Front. Comput. Sci. ›› 2026, Vol. 20 ›› Issue (2) : 2002315
DualMamba: a patch-based model with dual mamba for long-term time series forecasting
The field of time series forecasting has been seen widespread application of Transformer-based architectures. However, the quadratic complexity of the attention mechanism limits its performance in long-term time series forecasting. The proposition of patching mechanism has alleviated this issue to some extent, but models will struggle to effectively unify the information between intra-patch and inter-patch. To address this problem, we propose DualMamba, a novel Mamba-based model for time series forecasting, which segments the time series into subseries-level patches and employs dual Mamba modules to capture local and global information separately. Specifically, the time series use patch-wise dependencies to guide the local module, where each patch uses a point-wise representation of time series data. Furthermore, we design an information fusion mechanism for integrating information between intra-patch and inter-patch, which effectively incorporates global information into local contexts. This allows the model to capture both local details and global trends. Extensive experiments on several real-world datasets demonstrate that DualMamba achieves state-of-the-art performance in most cases and has reliable robustness, making it highly adaptable for various types of time series.
long-term time series forecasting / state space model / mamba / patching
| [1] |
|
| [2] |
|
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
Devlin J, Chang M W, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2019, 4171–4186 |
| [7] |
|
| [8] |
|
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
|
| [13] |
|
| [14] |
|
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
|
| [19] |
|
| [20] |
|
| [21] |
|
| [22] |
|
| [23] |
Wang Z, Kong F, Feng S, Wang M, Yang X, Zhao H, Wang D, Zhang Y. Is mamba effective for time series forecasting? Neurocomputing, 2025, 619: 129178 |
| [24] |
|
| [25] |
|
| [26] |
|
| [27] |
|
| [28] |
|
| [29] |
|
| [30] |
|
| [31] |
|
| [32] |
|
| [33] |
|
| [34] |
|
| [35] |
Das A, Kong W, Leach A, Mathur S, Sen R, Yu R. Long-term forecasting with tide: time-series dense encoder. Transactions on Machine Learning Research, 2023, 2023 |
| [36] |
Wu H, Hu T, Liu Y, Zhou H, Wang J, Long M. TimesNet: temporal 2D-variation modeling for general time series analysis. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [37] |
Liu M, Zeng A, Chen M, Xu Z, Lai Q, Ma L, Xu Q. SCINet: time series modeling and forecasting with sample convolution and interaction. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 421 |
| [38] |
Liu Y, Wu H, Wang J, Long M. Non-stationary transformers: Exploring the stationarity in time series forecasting. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 718 |
| [39] |
Wu H, Xu J, Wang J, Long M. Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. In: Proceedings of the 35th International Conference on Neural Information Processing Systems. 2021, 1717 |
Higher Education Press
/
| 〈 |
|
〉 |