Echo Flow Networks with Infinite-Horizon Memory

ICLR 2026 Conference Submission19295 Authors

19 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: reservoir computing; long-term memory; echo state networks; Transformer
TL;DR: We introduce Echo Flow Networks (EFNs), which extends classical Echo State Networks by modeling infinite memory.
Abstract: At the heart of time-series forecasting (TSF) lies a fundamental challenge: how can models efficiently and effectively capture long-range temporal dependencies across ever-growing sequences? While deep learning has brought notable progress, conventional architectures often face a trade-off between computational complexity and their ability to retain accumulative information over extended horizons. Echo State Networks (ESNs), a class of reservoir computing models, have recently regained attention for their exceptional efficiency, offering constant memory usage and per-step training complexity regardless of input length. This makes them particularly attractive for modeling extremely long-term event history in TSF. However, traditional ESNs fall short of state-of-the-art performance due to their limited nonlinear capacity, which constrains both their expressiveness and stability. We introduce ECHO FLOW NETWORKS (EFNS), a framework composed of a group of extended Echo State Networks (X-ESNs) with MLP readouts, enhanced by our novel Matrix-Gated Composite Random Activation (MCRA), which en- ables complex, neuron-specific temporal dynamics, significantly expanding the network’s representational capacity without compromising computational effi- ciency. In addition, we propose a dual-stream architecture in which recent input history dynamically selects signature reservoir features from an infinite-horizon memory, leading to improved prediction accuracy and long-term stability. Extensive evaluations on five benchmarks demonstrate that EFNS achieves up to 4× faster training and 3× smaller model size compared to leading methods like PatchTST, reducing forecasting error from 43% to 35%, a 20% relative improve- ment. One instantiation of our framework, EchoFormer, consistently achieves new state-of-the-art performance across five benchmark datasets: ETTh, ETTm, DMV, Weather, and Air Quality.
Supplementary Material: zip
Primary Area: learning on time series and dynamical systems
Submission Number: 19295
Loading