You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First , thanks for provide such a good graph calculation library。
Now I face a problem. When I use data of tens of thousands of nodes and millions of edges, Leiden's algorithm can complete community division in a few minutes on spark cluster (--spark.session.driverMemory=10g --spark.session.driverCores=1 --spark.session.executorCores=8 --spark.session.executorMemory=8G).
But when the data scale reaches one million nodes and 100 million edges, the algorithm will take 2 hours or more time.
So, for a large amount of data, is there any optimization method that can guide?
Looking forward to reply, thank you
The text was updated successfully, but these errors were encountered:
Thanks! Indeed, larger graphs may obviously take more time. The Leiden algorithm is one of the fastest algorithms available, so it won't be easy to find alternative. However, you might be interested in using the implementation in igraph itself: https://python.igraph.org/en/stable/api/igraph.Graph.html#community_leiden. That has more limited capabilities, but should be faster.
First , thanks for provide such a good graph calculation library。
Now I face a problem. When I use data of tens of thousands of nodes and millions of edges, Leiden's algorithm can complete community division in a few minutes on spark cluster (--spark.session.driverMemory=10g --spark.session.driverCores=1 --spark.session.executorCores=8 --spark.session.executorMemory=8G).
But when the data scale reaches one million nodes and 100 million edges, the algorithm will take 2 hours or more time.
So, for a large amount of data, is there any optimization method that can guide?
Looking forward to reply, thank you
The text was updated successfully, but these errors were encountered: