-
-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Metrics for 0.5 release #7059
Comments
@momack2 how important is getting metrics/benchmarks that assume the existence of hydra nodes? We can currently setup tests that are mixed network, Balsa-only and Cypress-only, however, we have yet to do one that includes any hydra-like behavior.
It might be useful to compare IPNS over PubSub resolution speed and update speeds to standard IPNS. However, given that that it's already existed in prior releases (albeit not independently from the DHT and less robust) if it will still be behind an experimental flag, which my understanding is that it will, then we may not need to collect these metrics now. Do we need to include metrics for FindPeer DHT requests? While it's not really "core IPFS" functionality (my understanding with no bearing whatsoever on project goals and direction) people do use |
All of our early metrics here are incorrect. They're helping us improve our query logic and helping us improve our metrics, but I don't expect the current metrics to look anything like the final metrics.
I believe the comments on hydra are simply calling out the fact that any metrics we get here won't fully reflect reality, but we should do our best. The core ask is:
The fact that this test won't include hydra is juts a caveat. |
@Stebalien is correct - core ask is repeatable metrics that don't include Hydra. SEPARATELY - Hydra should validate/prove through testing that these boosters improve network performance. @aschmahmann - I think those metrics would be really nice (so we have a datapoint on the performance gains of the new design), but lower priority than testing/releasing the DHT fixes. Aka, let's either defer until after the RC is cut, or delegate to another owner? If we expect swarm connect to be measurably different due to our DHT work and affect a common user flow, we should have a test/benchmark that exhibits the known change. |
Hey Folks - do we have an update on these metrics? @aschmahmann - is the testground graph you made still accurate or do we expect those estimates to change significantly? Would love to have a two different graph for public consumption (ex using labels like "0.5" instead of Cyprus) that shows benchmarks for each of finding & providing for nodes that upgrade (talking to other upgraded nodes vs old nodes) compared with current benchmarks of 0.4.23 and before. |
@momack2 I thought for simplicity we were only putting together graphs for testground tests a fully v0.4.23 network and for a fully v0.5.0 network, but not for a mixed network. We have a testground test ready to go for a mixed network, but @jacobheun thought the nuances and assumptions in the tests might be a bit more complicated to explain. The TLDR is that performance gets better for everyone once we switch the network over to the new protocol version, but go-ipfs v0.5.0 nodes have a larger performance benefit. Do we want graphs for a mixed network? |
Let's stick with a simple "old DHT" versus "new DHT" graph. Depending on the network load and hydra performance, network performance for "new" nodes may increase or may decrease as the rest of the network upgrades. Case 1: Increase: Hydra is not powerful enough to easily handle requests on the network. As more nodes join the network, they share the load. Case 2: Decrease: Hydra is powerful enough to handle all requests on the network. As more nodes join the network, they'll likely perform worse than the hydra nodes, slowing down requests. |
My bad - yes, fine to just look at a totally upgraded network (I was staring at your previous graph that also had metrics for cyprus searching balsa when I asked) |
We need metrics for the 0.5 release to give IPFS current and potential users an estimate of the performance improvements they should see in this release from the various improvements included. These metrics should be as realistic and repeatable as possible since we are continuing to merge additional patches and improvements over time. This is particularly hard to do since combining features together (DHT changes + hydra + "% of network upgraded") produce very different metrics. Therefore, I suggest we mostly narrow in on "fully upgraded" network scenarios to limit the complexity of our testing.
I'd like us to define repeatable tests to validate/benchmark the major performance improvements we see in this release:
Other useful tests:
@alanshaw @aschmahmann @Stebalien @petar - are there other major aspects we need to benchmark or include in these tests?
The text was updated successfully, but these errors were encountered: