Scalable Hypergraph Learning and Processing
Citations Over TimeTop 20% of 2015 papers
Abstract
A hypergraph allows a hyperedge to connect more than two vertices, using which to capture the high-order relationships, many hypergraph learning algorithms are shown highly effective in various applications. When learning large hypergraphs, converting them to graphs to employ the distributed graph frameworks is a common approach, yet it results in major efficiency drawbacks including an inflated problem size, the excessive replicas, and the unbalanced workloads. To avoid such drawbacks, we take a different approach and propose HyperX, which is a thin layer built upon Spark. To preserve the problem size, HyperX directly operates on a distributed hypergraph. To reduce the replicas, HyperX replicates the vertices but not the hyperedges. To balance the workloads, we investigate the hypergraph partitioning problem aiming at minimizing the space and the communication cost subject to two separate constraints on the hyperedge and the vertex workloads. With experiments on both real and synthetic datasets, we verify that HyperX significantly improves the efficiency of the learning algorithms when compared with the graph conversion approach.
Related Papers
- → Hypergraph Modeling(2023)1 cited
- → Decompositions of 3-uniform hypergraph K_v^{(3)} into hypergraph K_4^{(3)}+e(2010)
- → On the Random Greedy $F$-Free Hypergraph Process(2015)
- → Non-uniform Hypergraphs(2020)
- On the random greedy F-free hypergraph process(2015)