Recognition: unknown
Predicting Channel Closures in the Lightning Network with Machine Learning
Pith reviewed 2026-05-14 21:09 UTC · model grok-4.3
The pith
Temporal and behavioral signals from public gossip data predict Lightning Network channel closures, while network topology adds no value.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
We construct a dataset spanning over two years of LN activity and benchmark a range of machine learning approaches, from MLPs to temporal graph neural networks and spectral encodings. Our experiments reveal that the dominant predictive signals are temporal and behavioural, namely how recently each endpoint was active and the per-node history of past closures, while the surrounding network topology provides no additional benefit. We find that a simple MLP operating on edge-level features, node-level event counts, and temporal patterns outperforms all graph-based approaches.
What carries the argument
Temporal link classification over the evolving channel graph, using edge-level features, node event counts, and temporal patterns as input to an MLP.
If this is right
- Network participants can anticipate forced closures using only recent activity history and node-level patterns.
- Graph-based models are unnecessary for this prediction task since topology provides no lift.
- The inherent privacy of balances and flows sets a fundamental limit on closure predictability from gossip data alone.
- Releasing the dataset enables further work on practical reliability improvements in payment channel networks.
Where Pith is reading between the lines
- Designers of future layer-two networks might consider exposing limited additional signals if higher closure predictability is a goal.
- The same temporal-feature approach could transfer to other dynamic graphs where node behavior drives edge events.
- Individual node history likely captures local incentives that global structure does not, suggesting closures are mostly local decisions.
Load-bearing premise
Publicly available gossip data contains sufficient temporal and behavioral signals to predict closure types despite the privacy of channel balances and payment flows.
What would settle it
Training a graph neural network on the full topology and node features and finding that it achieves materially higher accuracy than the MLP on the held-out portion of the two-year dataset.
Figures
read the original abstract
The Lightning Network (LN) is a second-layer protocol for Bitcoin designed to enable fast and cost-efficient off-chain transactions. Channels in the LN can be closed either by mutual agreement or unilaterally through a forced closure, which locks the involved capital for an extended period and degrades network reliability. In this paper, we study the problem of predicting channel closure types from publicly available gossip data, framing it as a temporal link classification task over the evolving channel graph. We construct a dataset spanning over two years of LN activity and benchmark a range of machine learning approaches, from MLPs to temporal graph neural networks and spectral encodings. Our experiments reveal that the dominant predictive signals are temporal and behavioural, namely how recently each endpoint was active and the per-node history of past closures, while the surrounding network topology provides no additional benefit. We find that a simple MLP operating on edge-level features, node-level event counts, and temporal patterns outperforms all graph-based approaches, and discuss how the inherent privacy of the LN, where critical information such as channel balances and payment flows remains hidden, fundamentally limits the predictability of closures from gossip data alone. We publicly release the dataset and code at https://github.com/AmbossTech/ln-channel-closure-prediction to encourage further research on this practically relevant task.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper frames channel closure prediction in the Lightning Network as a temporal link classification task on gossip data spanning over two years. It benchmarks MLPs against temporal GNNs and spectral encodings, claiming that temporal and behavioral signals (recent activity and per-node closure history) dominate while network topology adds no value, with a simple MLP outperforming graph-based methods. The work releases the dataset and code.
Significance. If the central empirical finding holds after proper controls, the result usefully documents the practical limits of public gossip data for LN closure prediction due to hidden balances and flows, and supplies a reproducible benchmark dataset for temporal link prediction in payment networks.
major comments (3)
- [§4] §4 (Experiments) and Table 2: the reported superiority of the MLP over temporal GNNs and spectral methods lacks any description of the hyperparameter search budget, depth of message passing, temporal aggregation windows, or number of independent runs with error bars; without these the performance gap cannot be confidently attributed to absence of topological signal rather than implementation choices.
- [§3.2] §3.2 (Feature construction): node-level event counts and temporal patterns are described at a high level but the exact time-windowing, normalization, and handling of the dynamic edge set over the two-year span are not specified, making it impossible to assess whether the temporal features already encode the limited topological information available in gossip data.
- [§4.3] §4.3 (Validation): the manuscript provides no cross-validation scheme, train/test temporal split details, or full metric suite (precision, recall, F1, AUC) with confidence intervals; the abstract claim that temporal features dominate therefore rests on moderate rather than strong empirical support.
minor comments (2)
- [Abstract] The GitHub link is given but the repository structure and exact data preprocessing scripts are not described in the text, which would aid reproducibility.
- [Figure 1] Figure 1 (network evolution) would benefit from clearer labeling of the time axis and closure-type color coding.
Simulated Author's Rebuttal
We thank the referee for the constructive comments on our manuscript. We have revised the paper to incorporate additional methodological details as requested, which we believe strengthens the empirical claims without altering the core findings.
read point-by-point responses
-
Referee: [§4] §4 (Experiments) and Table 2: the reported superiority of the MLP over temporal GNNs and spectral methods lacks any description of the hyperparameter search budget, depth of message passing, temporal aggregation windows, or number of independent runs with error bars; without these the performance gap cannot be confidently attributed to absence of topological signal rather than implementation choices.
Authors: We agree that these details are necessary for rigorous interpretation. In the revised manuscript we have expanded Section 4 to describe the hyperparameter search (grid search over learning rates [1e-4, 1e-2], hidden dimensions [32, 128], GNN layers [1, 3], and temporal windows of 1/7/30 days), the message-passing depth used for the temporal GNN baselines, and the fact that all results in Table 2 are means and standard deviations over five independent runs with different random seeds. The performance advantage of the MLP remains consistent across these settings. revision: yes
-
Referee: [§3.2] §3.2 (Feature construction): node-level event counts and temporal patterns are described at a high level but the exact time-windowing, normalization, and handling of the dynamic edge set over the two-year span are not specified, making it impossible to assess whether the temporal features already encode the limited topological information available in gossip data.
Authors: We have revised Section 3.2 to provide the exact specifications: node-level features are computed over three fixed sliding windows (1 day, 7 days, 30 days) ending at the snapshot time; counts are normalized by the total events observed in each window across the entire network; the dynamic edge set is handled by restricting all features to channels that are active (i.e., have not yet closed) at the prediction timestamp, with no future information leaked. These choices are now stated explicitly so that readers can evaluate whether topological signal is already captured by the temporal aggregates. revision: yes
-
Referee: [§4.3] §4.3 (Validation): the manuscript provides no cross-validation scheme, train/test temporal split details, or full metric suite (precision, recall, F1, AUC) with confidence intervals; the abstract claim that temporal features dominate therefore rests on moderate rather than strong empirical support.
Authors: We accept this point. The revised Section 4.3 now specifies a strict temporal train/test split (first 18 months for training, final 6 months for testing) to preserve causality, explains why k-fold cross-validation is inappropriate for this time-series setting, and reports the full metric suite (precision, recall, F1, AUC) together with 95 % confidence intervals obtained from the five independent runs. These additions provide stronger quantitative support for the dominance of temporal and behavioral features. revision: yes
Circularity Check
No significant circularity in empirical ML evaluation
full rationale
The paper presents an empirical machine learning study: it constructs a two-year LN gossip dataset, frames closure prediction as temporal link classification, and benchmarks MLPs against temporal GNNs and spectral methods. All central claims (temporal/behavioral signals dominate, topology adds no benefit, MLP outperforms graph models) are stated as direct experimental outcomes on the released data and code. No equations, first-principles derivations, self-definitional parameters, or load-bearing self-citations exist that would reduce any reported prediction to its own inputs by construction. The work is self-contained against external benchmarks and independently reproducible, yielding a normal finding of zero circularity.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
The bitcoin lightning network: Scalable off-chain instant payments,
Poon and Dryja, “The bitcoin lightning network: Scalable off-chain instant payments,” 2016
2016
-
[2]
Node classification and geographical analysis of the lightning cryptocurrency network,
P. Zabka, K.-T. F ¨orster, S. Schmid, and C. Decker, “Node classification and geographical analysis of the lightning cryptocurrency network,” inProceedings of the 22nd International Conference on Distributed Computing and Networking, ser. ICDCN ’21. New York, NY , USA: Association for Computing Machinery, 2021, p. 126–135. [Online]. Available: https://doi...
-
[3]
On the difficulty of hiding the balance of lightning network channels,
J. Herrera-Joancomart ´ı, G. Navarro-Arribas, A. Ranchal-Pedrosa, C. P ´erez-Sol`a, and J. Garcia-Alfaro, “On the difficulty of hiding the balance of lightning network channels,” inProceedings of the 2019 ACM Asia Conference on Computer and Communications Security, ser. Asia CCS ’19. New York, NY , USA: Association for Computing Machinery, 2019, p. 602–61...
-
[4]
Benchmarking gnns using lightning network data,
R. Feichtinger, F. Gr ¨otschla, L. Heimbach, and R. Wattenhofer, “Benchmarking gnns using lightning network data,” 2024. [Online]. Available: https://arxiv.org/abs/2407.07916
-
[5]
Aravind Sankar, Yanhong Wu, Liang Gou, Wei Zhang, and Hao Yang
E. Rossi, B. Chamberlain, F. Frasca, D. Eynard, F. Monti, and M. M. Bronstein, “Temporal graph networks for deep learning on dynamic graphs,”CoRR, vol. abs/2006.10637, 2020. [Online]. Available: https://arxiv.org/abs/2006.10637
-
[6]
Temporal graph benchmark for machine learning on temporal graphs,
S. Huang, F. Poursafaei, J. Danovitch, M. Fey, W. Hu, E. Rossi, J. Leskovec, M. M. Bronstein, G. Rabusseau, and R. Rabbany, “Temporal graph benchmark for machine learning on temporal graphs,” inThirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2023. [Online]. Available: https://openreview.net/forum?id=qG7IkQ7IBO
2023
-
[7]
TGB 2.0: A benchmark for learning on temporal knowledge graphs and heterogeneous graphs,
J. Gastinger, S. Huang, M. Galkin, E. Loghmani, A. Parviz, F. Poursafaei, J. Danovitch, E. Rossi, I. Koutis, H. Stuckenschmidt, R. Rabbany, and G. Rabusseau, “TGB 2.0: A benchmark for learning on temporal knowledge graphs and heterogeneous graphs,” inThe Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2024. ...
2024
-
[8]
T. Chen and C. Guestrin, “Xgboost: A scalable tree boosting system,” inProceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ser. KDD ’16. New York, NY , USA: Association for Computing Machinery, 2016, p. 785–794. [Online]. Available: https://doi.org/10.1145/2939672.2939785
-
[9]
Lightgbm: A highly efficient gradient boosting decision tree,
G. Ke, Q. Meng, T. Finley, T. Wang, W. Chen, W. Ma, Q. Ye, and T.-Y . Liu, “Lightgbm: A highly efficient gradient boosting decision tree,” inAdvances in Neural Information Processing Systems, I. Guyon, U. V . Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds., vol. 30. Curran Associates, Inc., 2017. [Online]. Available: https...
2017
-
[10]
Inductive representation learning on large graphs,
W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” inAdvances in Neural Information Processing Systems, I. Guyon, U. V . Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds., vol. 30. Curran Associates, Inc., 2017. [Online]. Available: https://proceedings.neurips.cc/paper files/paper/201...
2017
-
[11]
Topological analysis of bitcoin’s lightning network,
I. A. Seres, L. Guly ´as, D. A. Nagy, and P. Burcsi, “Topological analysis of bitcoin’s lightning network,” inMathematical Research for Blockchain Economy, P. Pardalos, I. Kotsireas, Y . Guo, and W. Knot- tenbelt, Eds. Cham: Springer International Publishing, 2020, pp. 1–12
2020
-
[12]
Channel balance interpolation in the lightning network via machine learning,
E. Rossi, V . Singhet al., “Channel balance interpolation in the lightning network via machine learning,”arXiv preprint arXiv:2405.12087, 2024
-
[13]
M. Salahshour, A. Shafiee, and M. Tefagh, “Joint combinatorial node selection and resource allocations in the lightning network using attention-based reinforcement learning,” 2024. [Online]. Available: https://arxiv.org/abs/2411.17353
-
[14]
V . Singh, M. Khanzadeh, V . Davis, H. Rush, E. Rossi, J. Shrader, and P. Lio, “Bayesian binary search,”arXiv preprint arXiv:2410.01771, 2024
-
[15]
Evolvegcn: Evolving graph convolutional networks for dynamic graphs,
A. Pareja, G. Domeniconi, J. Chen, T. Ma, T. Suzumura, H. Kanezashi, T. Kaler, T. Schardl, and C. Leiserson, “Evolvegcn: Evolving graph convolutional networks for dynamic graphs,”Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, pp. 5363–5370, Apr. 2020. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/ article/view/5984
2020
-
[16]
Gc-lstm: graph convolution embedded lstm for dynamic network link prediction,
J. Chen, X. Wang, and X. Xu, “Gc-lstm: graph convolution embedded lstm for dynamic network link prediction,”Applied Intelligence, vol. 52, no. 7, p. 7513–7528, May 2022. [Online]. Available: https://doi.org/10.1007/s10489-021-02518-9
-
[17]
Discrete- time temporal network embedding via implicit hierarchical learning in hyperbolic space,
M. Yang, M. Zhou, M. Kalander, Z. Huang, and I. King, “Discrete- time temporal network embedding via implicit hierarchical learning in hyperbolic space,” inProceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, ser. KDD ’21. New York, NY , USA: Association for Computing Machinery, 2021, p. 1975–1985. [Online]. Available: https...
-
[18]
Roland: Graph learning framework for dynamic graphs,
J. You, T. Du, and J. Leskovec, “Roland: Graph learning framework for dynamic graphs,” inProceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, ser. KDD ’22. New York, NY , USA: Association for Computing Machinery, 2022, p. 2358–2366. [Online]. Available: https://doi.org/10.1145/3534678.3539300
-
[19]
Wingnn: Dynamic graph neural networks with random gradient aggregation window,
Y . Zhu, F. Cong, D. Zhang, W. Gong, Q. Lin, W. Feng, Y . Dong, and J. Tang, “Wingnn: Dynamic graph neural networks with random gradient aggregation window,” inProceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, ser. KDD ’23. New York, NY , USA: Association for Computing Machinery, 2023, p. 3650–3662. [Online]. Available:...
-
[20]
Dyrep: Learning representations over dynamic graphs,
R. Trivedi, M. Farajtabar, P. Biswal, and H. Zha, “Dyrep: Learning representations over dynamic graphs,” inInternational Conference on Learning Representations, 2019. [Online]. Available: https: //openreview.net/forum?id=HyePrhR5KX
2019
-
[21]
Predicting dynamic embedding trajectory in temporal interaction networks,
S. Kumar, X. Zhang, and J. Leskovec, “Predicting dynamic embedding trajectory in temporal interaction networks,” inProceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ser. KDD ’19. New York, NY , USA: Association for Computing Machinery, 2019, p. 1269–1278. [Online]. Available: https://doi.org/10.1145/3292500.3330895
-
[22]
Inductive representation learning on temporal graphs,
da Xu, chuanwei ruan, evren korpeoglu, sushant kumar, and kannan achan, “Inductive representation learning on temporal graphs,” in International Conference on Learning Representations, 2020. [Online]. Available: https://openreview.net/forum?id=rJeW1yHYwH
2020
-
[23]
Towards better dynamic graph learning: New architecture and unified library,
L. Yu, L. Sun, B. Du, and W. Lv, “Towards better dynamic graph learning: New architecture and unified library,” inThirty-seventh Conference on Neural Information Processing Systems, 2023. [Online]. Available: https://openreview.net/forum?id=xHNzWHbklj
2023
-
[24]
Do we really need complicated model architectures for temporal networks?
W. Cong, S. Zhang, J. Kang, B. Yuan, H. Wu, X. Zhou, H. Tong, and M. Mahdavi, “Do we really need complicated model architectures for temporal networks?” inThe Eleventh International Conference on Learning Representations, 2023. [Online]. Available: https://openreview.net/forum?id=ayPPc0SyLv1
2023
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.