I have doubts that advanced pathfinding techniques are even that useful in a highly malleable network like LN.
Pathfinding is huge in static networks like roads (traveling salesman, etc.)
But in the LN, if there is a long path from point A to point B, you can just make a new path (a.k.a. channel) direct to the destination and for pretty cheap and have it operational in less than an hour.
reply
Our benchmark testing is ongoing and the initial results are extremely promising vs off the shelf LND.
In a dynamic graph like Lightning, the need for advanced computation is even greater.
reply
By "off the shelf LND", do you mean a new LND node that has not yet built a robust probabilistic routing model?
LND uses probabilistic routing to build a local model of payment success probabilities for every attempted path. After many payment attempts, an LND node should optimize its own pathfinding.
reply
We're testing both zeroed-out mission control and a learned mission control.
Apriori is decidedly difficult to benchmark against!
reply
I'm suspicious of machine learning path finding in general. Big fan of machine learning. But classical algorithms are already good in these graph algorithms
reply
Totally agree with this... what to benchmark will be an issue and it's liquid (highly malleable as you said)... I don't know if the effort will pay off, too much computation to adapt every change in the graph.
reply
The computation and retraining is not intensive at this size of the network graph. As the network graph increases in size, the improvements in payment reliability offered by this type of calculation should increase versus anarchistic, source-based routing.
reply
BOYCOTT AMBOSS! YOU HAVE BEEN WARNED.
reply
  • how many nodes and channels were in the dataset used to train this?
  • i think it is basically impossible to do a proper train/test/validation here. for example, if you have one side of the channel in the training set then the other side might be in the test set and you have leaked the answer to your model to learn. you can't fully sequester this information. how did you try to deal with this?
reply
reply
That leaks node-level features from the train set into the test/validation set then no? Like if 90% of my channels are in the train set, and the top performing feature is "positional encoding" then the train set can basically learn my nodes positional encoding and whether my channels tend to be balanced or not. Then it just extrapolates that to my channels in the test/validation set.
reply
Any effort to solve scalability is a great opportunity.
reply
deleted by author
reply