WormHole is a graph algorithm that outperforms traditional traversal and indexing-based methods like BiBFS, PLL, and MLL. Through extensive benchmarking on large datasets from SNAP and KONECT, WormHole demonstrates superior efficiency in query cost, inquiry time, and setup performance. It achieves 99% path accuracy with minimal overhead, offering faster and more scalable solutions for large graph queries—particularly when compared to indexing algorithms that time out or require massive resources.WormHole is a graph algorithm that outperforms traditional traversal and indexing-based methods like BiBFS, PLL, and MLL. Through extensive benchmarking on large datasets from SNAP and KONECT, WormHole demonstrates superior efficiency in query cost, inquiry time, and setup performance. It achieves 99% path accuracy with minimal overhead, offering faster and more scalable solutions for large graph queries—particularly when compared to indexing algorithms that time out or require massive resources.

WormHole Algorithm Outperforms BiBFS in Query Efficiency and Accuracy

2025/10/16 21:00

Abstract and 1. Introduction

1.1 Our Contribution

1.2 Setting

1.3 The algorithm

  1. Related Work

  2. Algorithm

    3.1 The Structural Decomposition Phase

    3.2 The Routing Phase

    3.3 Variants of WormHole

  3. Theoretical Analysis

    4.1 Preliminaries

    4.2 Sublinearity of Inner Ring

    4.3 Approximation Error

    4.4 Query Complexity

  4. Experimental Results

    5.1 WormHole𝐸, WormHole𝐻 and BiBFS

    5.2 Comparison with index-based methods

    5.3 WormHole as a primitive: WormHole𝑀

References

5 EXPERIMENTAL RESULTS

In this section, we experimentally evaluate the performance of our algorithm. We look at several metrics to evaluate performance in different aspects. We compare with BiBFS, a traversal-based approach, and with the indexing algorithms PLL and MLL. We test several aspects, summarized next. Detailed results are provided in the rest of this section.

\ (1) Query cost: By query cost, we refer to the number of vertices queried by WormHole, consistent with our access model (see §1.2). We show that WormHole actually does remarkably well in terms of query cost, seeing a small fraction of the whole graph even for several thousands of shortest path inquiries. See Figures 2(b) and 5.

\ (2) Inquiry time: We demonstrate that WormHole𝐸 achieves consistent speedups over traditional BiBFS, even while using it as the sole primitive in the procedure. More complex methods such as PLL and MLL time out for the majority of large graphs. We also provide variants that achieve substantially higher speedups. Finally, in §5.3, we show how using the existing indexing-based state or the art methods on the core lets us achieve indexing-level inquiry times. See Figure 1.

\ (3) Accuracy: We show that our estimated shortest paths are accurate up to an additive error 2 on 99% of the inquiries for the default version WormHole𝐸; a faster heuristic, WormHole𝐻 , shows lower accuracy, but still over 90% of inquiries satisfy this condition. See §5.1 and Table 3 for details.

\ (4) Setup: We look at the setup time and disk space with each associated method. Perhaps as expected, WormHole𝐸 beats the indexing based algorithms by a wide margin in terms of both space and time: see Figure 7. In §5.3 we further show that using these methods restricted to Cin results in a variant WormHole𝑀 with much lower setup cost ( Table 6).

\ Datasets. The experiments have been carried out on a series of datasets of varying sizes, as detailed in Table 2. The datasets have been taken either from the SNAP large networks database [36] or the KONECT project [34]. We organize the results into two broad sections: we first introduce two variants

\

\ Table 2: Network datasets used for experimental evaluation with their corresponding sizes. We observe that BiBFS finishes on all the datasets, but the indexing based methods do not on the medium and large networks. We were able to set up MLL on large-dblp in reasonable time, but the subsequent shortest path inquiries were met with consistent segmentation faults that we were unable to debug.

\ Table 3: Summary of WormHole with the two cases: WormHole𝐸, with the exact shortest path through the inner ring, and WormHole𝐻 that picks only the shortest path between the highest degree vertices – refer to §5.1. We note the mean inquiry times per inquiry (MIT) in microseconds, and average speed up per inquiry (SU/I) compared to BiBFS for each method. We also note the percentiles of inquiries by absolute error: for WormHole𝐸, we get absolute error under 2 for over 99% of the inquiries. This drops for WormHole𝐻 , but it is still above 99% for six of the ten datasets, and over 90% in all of them. Accuracy numbers are highlighted in green, where darker is better. Similarly, we have a gradient of violet for speedups; darker is faster. For WormHole𝐸, speedup over BiBFS per inquiry on average is usually between 2× and 3×, but this increases to consistently between 20−30× in WormHole𝐻 , and reaches a max of 181× in our largest dataset, soc-twitter.

\ of our algorithm. We then compare it with BiBFS as well as indexing based methods – PLL and MLL. The latter two did not terminate in 12 hours for most of the graphs, while BiBFS completed on even our largest networks.

\ We classify the examined graphs into three different classes and use a fixed percentage as the ‘optimal’ inner ring size for graphs of comparable size (where the inner core size as % of the total size decreases for larger networks, an indication for the sublinearity of our approach). This takes into account the tradeoff between accuracy and the query/memory costs incurred by a larger inner ring. The classification is summarized in Table 1. For the experimental section, we default to these sizes unless mentioned otherwise.

\ Implementation details. We run our experiments on an AWS ec2 instance with 32 AMD EPYC™ 7R32 vCPUs and 64GB of RAM. The code is written in C++ and is available in the supplementary material as a zipped folder, with links to the datasets. The backbone of the graph algorithms is a subgraph counting library that uses compressed sparse representations [45].

\

:::info Authors:

(1) Talya Eden, Bar-Ilan University (talyaa01@gmail.com);

(2) Omri Ben-Eliezer, MIT (omrib@mit.edu);

(3) C. Seshadhri, UC Santa Cruz (sesh@ucsc.edu).

:::


:::info This paper is available on arxiv under CC BY 4.0 license.

:::

\

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.
Share Insights

You May Also Like

Ethereum's "double crisis": core talent continues to leave, and technical debt quietly accumulates

Ethereum's "double crisis": core talent continues to leave, and technical debt quietly accumulates

By Eric, Foresight News On the evening of the 19th Beijing time, Bankless co-founder David Hoffman posted a message on X to "mourn" Dankrad Feist, the longest-serving researcher at the Ethereum Foundation, who chose to leave Ethereum and join the stablecoin L1 Tempo. David Hoffman believes the issue of for-profit companies co-opting the talent cultivated by the Ethereum open-source community is significant, and argues that these companies do not, as they claim, bring greater benefits to Ethereum. He bluntly stated, "In my view, Tempo's purpose is to intercept the trillions of dollars in stablecoins expected to flow in over the next decade and place them on their private blockchain. While this will certainly expand the market, Tempo still intends to grab as much of the pie as possible." He believes Tempo will inevitably be constrained by compliance issues, which even issuing tokens cannot address. While both Tempo and Ethereum will bring change to the world, Ethereum is uniquely suited to serve as a trusted, neutral global settlement layer, without shareholders and unconstrained by law. The feeling of disappointment with Ethereum began to surface when its price began to lag behind Bitcoin's in this cycle. However, over time, people began to realize that the exodus of talented individuals from the Ethereum community seemed irreversible. When dreams conflicted with self-interest, many ultimately chose the latter, a fact that many in the industry have long worried about. Dankrad Feist is not the first and will not be the last Dankrad Feist announced his joining Tempo at X on the 17th of this month and stated that he would continue to serve as a research advisor for the Ethereum Foundation's Protocol Cluster's three strategic initiatives: scaling Layer 1, scaling Blobs, and improving user experience. He stated, "Ethereum has strong values and technology choices that make it unique. Tempo will be a great complement, building on similar technology and values while pushing boundaries in scale and speed. I believe this will be a significant benefit to Ethereum. Tempo's open-source technology can be easily integrated back into Ethereum, benefiting the entire ecosystem." According to LinkedIn, Dankrad Feist officially joined Ethereum as a researcher in 2019, focusing on sharding technology, which can scale the Ethereum mainnet. Danksharding, one of the core components of Ethereum's current scaling roadmap, is named after him. Danksharding is a key technical path for Ethereum to achieve high-throughput and low-cost transactions, and is widely considered by the community to be the most important upgrade direction after Ethereum 2.0. Dankrad Feist promoted Proto-Danksharding (EIP-4844), a predecessor of Danksharding. This EIP introduced the blob transaction type, providing a cheaper and more efficient data availability layer for Rollup, significantly reducing the data publishing cost of Rollup. In addition, he had a public debate with Geth development lead Péter Szilágyi on the MEV issue, which eventually prompted Vitalik to step in to coordinate and promote the community's attention to MEV mitigation mechanisms (such as PBS, Proposer-Builder Separation). Tempo researcher Mallesh Pai introduced the members joining Tempo in September, and Liam Horne, former CEO of OP Labs and co-founder of ETHGlobal, also appeared on the list. Before Dankrad Feist, the person who surprised the industry was Danny Ryan, who co-founded Etherealize, a $40 million funding round. A former core member of the Ethereum Foundation and known as the "Chief Engineer of Ethereum 2.0," Ryan joined Etherealize just six months after announcing his indefinite departure in September 2024. However, given that Etherealize shares similarities with ConsenSys, founded by Ethereum co-founder Joseph Lubin 11 years prior amidst controversy over commercialization, Ryan's departure has been widely understood. What really worries David Hoffman are companies like Tempo and Paradigm. Well-known Ethereum developer Federico Carrone expressed a similar sentiment, retweeting David Hoffman's tweet about Dankrad Feist joining Tempo and stating that he has been saying for the past two years that Paradigm's influence within Ethereum could become a tail risk for the entire ecosystem. Federico Carrone wrote that the sole goal of a venture capital fund is to maximize returns for its limited partners. Ethereum shouldn't become deeply dependent on the technology of a venture capital firm that is playing its cards with extreme strategic skill. Following the FTX debacle, Paradigm removed nearly all cryptocurrency-related branding and made a high-profile shift to AI. Carrone believes this is proof enough of his point. After Trump returned to the White House, Paradigm re-entered the Web3 space, aggressively recruiting top researchers from the community, funding key Ethereum open-source libraries, and supporting Stripe's launch of Tempo. Carrone believes that while Paradigm claims its work is beneficial to Ethereum—more funding, more tools, more testing grounds, and the potential for new ideas to feed back into Ethereum—are all potential benefits, but when corporations have excessive visibility and influence over open-source projects, priorities shift from the community's long-term vision to corporate profits. Ethereum’s technical debt is accumulating The simple loss of talent in the Ethereum open source community may not cause widespread concern, but if the loss of talent is accompanied by the accumulation of technical debt, it is worthy of high vigilance. A week ago, a community user posted a screenshot on X, revealing that Solidity's top contributors have all but ceased development. Only Cameel continues to raise new issues and advance the technology, but appears to be in maintenance mode. He believes the community needs to invest more resources in supporting the programming language. Some users in the comments questioned why efforts were being expended on continuously improving and upgrading Solidity rather than simply maintaining it to ensure stability and security. The user who tweeted explained that even changing the Solidity compiler wouldn't change any deployed contracts, but could improve security, enhance the development experience, or support the use of new contracts. As can be seen in the chart above, development activity began to decline sharply at the beginning of the previous bull market. Federico Carrone also expressed his concern, stating that his biggest concern is that the numerous core tools and libraries built around Solidity may not receive long-term maintenance. Even the latest Solidity compiler is currently supported by only a handful of developers. Furthermore, companies involved in L2 and ZK technologies are downsizing, leaving the final iteration of cutting-edge technologies to a handful of companies. With increasing gas limits, many execution clients have not seen substantial performance improvements, and judging by the libraries, the development teams of these clients appear to be lagging behind. Federico Carrone said, “Ethereum’s technical debt continues to accumulate, not only because the protocol itself must continue to evolve, but also because many of its dependencies and surrounding repositories have become stagnant. The entire ecosystem continues to expand, protecting tens of billions of dollars in assets, while part of its foundation is quietly eroding.” Open source communities cannot simply "generate power with love" For an open-source community like Ethereum, which carries a vast amount of value that can be measured in real money, balancing "generate power with love" and economic incentives is a problem without any real precedent. This should be a matter of great concern to the Ethereum Foundation, but it seems to have been overlooked. Péter Szilágyi, who joined the Ethereum Foundation in 2015 and is responsible for the development and maintenance of Geth, clearly pointed out the three most disappointing problems in a letter to the leadership of the Ethereum Foundation a year and a half ago: being portrayed as a leader externally but marginalized internally; the serious disproportion between income and the growth of Ethereum's market value; and Vitalik and a small group of people around him having too much say in the Ethereum ecosystem. In late 2024, Péter Szilágyi discovered that the Ethereum Foundation was secretly incubating an independent fork of Geth. He was subsequently fired due to a dispute with the Ethereum Foundation and repeatedly declined rehire. The Ethereum Foundation even offered Szilágyi $5 million to separate Geth from the Foundation, but was rejected. Currently, Szilágyi maintains the Geth codebase as an independent contributor. Rumors of corruption within the Ethereum Foundation have been circulating, but this is a problem that should have been anticipated from the moment the Ethereum Foundation was founded. As the saying goes, "where there are people, there are gangs." We can't eliminate human greed, but we also can't allow Ethereum to gradually lose its core value due to commercialization. Ethereum's market capitalization of hundreds of billions of dollars, having handled trillions of dollars in on-chain value transfers for years, is built on infrastructure built by a professional technical team, centered on a permissionless, open-source ethos, and commercialized by a large number of businesses. However, simply maintaining such a massive system requires a significant workforce, and as we've discussed, these individuals are leaving due to disappointment or opting for other projects driven by financial gain. The Ethereum Foundation underwent drastic reforms this year, but so far, they haven't produced any significant results. Ethereum can still be called the world's computer, and its potential for commercial applications is still being explored by talented teams. However, as the foundation of all this, Ethereum cannot continue to disappoint those who still hold on to its ideals.
Share
2025/10/23 09:01
Share