The content discusses the optimization of the Leiden algorithm for community detection in shared memory settings. It introduces GVE-Leiden, which achieves high processing rates and improved performance compared to other implementations like the original Leiden, igraph Leiden, and NetworKit Leiden.
Community detection is crucial in various applications with large datasets. The Louvain method may produce disconnected communities, leading to the proposal of the Leiden algorithm by Traag et al. The refinement phase in the Leiden algorithm allows for better identification of well-connected communities.
Optimization techniques are applied to enhance the aggregation phase of the Leiden algorithm. Various strategies such as dynamic loop scheduling and threshold scaling are utilized to improve efficiency. Results show that a greedy approach performs best in terms of runtime and modularity.
The experimental setup includes a server with dual Intel Xeon Gold 6226R processors and graphs sourced from SuiteSparse Matrix Collection. Performance comparisons show that GVE-Leiden outperforms other implementations in terms of runtime, processing rate, and modularity.
Disconnected communities are identified using a parallel algorithm that explores different approaches involving BFS or DFS traversal methods. The evaluation demonstrates the effectiveness of GVE-Leiden in achieving faster processing rates and improved performance compared to existing implementations.
Na inny język
z treści źródłowej
arxiv.org
Głębsze pytania