-
Notifications
You must be signed in to change notification settings - Fork 445
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Performance issue using abortableSource() #1420
Comments
A core value in Rust is "zero cost abstractions". As Javascript developers we must be mindful that in this language this is not the case. I worry that in the current libp2p implementation a single package requires a very high amount of iterator cycles, just for a single package. More investigation is needed to understand the costs, but on a preliminary review it's a apparent that the amount of Promises and related objects per packet may be excessive. |
Javascript abstractions are not free. While using pipe here looks nice, it adds a non-neglible cost allocating Promises for each extra `for await ()` iteration. - Similar rationale to libp2p/js-libp2p#1420 (comment) This PR merges sink pipe components in a single iteration Co-authored-by: achingbrain <[email protected]>
## [7.0.6](v7.0.5...v7.0.6) (2022-11-25) ### Bug Fixes * reduce async iterator loops per package in _createSink ([#224](#224)) ([e2a32ad](e2a32ad)), closes [/github.com/libp2p/js-libp2p/issues/1420#issuecomment-1273272662](https://github.com/libp2p//github.com/libp2p/js-libp2p/issues/1420/issues/issuecomment-1273272662)
Usage of `abortable-iterator` has a non-negligible cost, see libp2p/js-libp2p#1420 so remove it and close the socket directly when the abort signal fires. Co-authored-by: achingbrain <[email protected]>
Did some investigation. In the await Promise.race([iterator.next(), abortPromise]) If you, instead do it in batches of 50-100, like this, await Promise.race([Promise.all(batch), abortPromise]) where batch is an array of iteration promises, you get abourt 50% improvement in performance, but you will have a promise that is running the background until the batch is fully resolved or rejected (even if iterator is aborted) Even faster, if you consider that we necessarily don't need to cancel promises att all and do something like async function* asyncIterableWithAbort<T>(signal: AbortSignal, iterable: AsyncIterable<T>): AsyncGenerator<T, void, undefined> {
const iterator = iterable[Symbol.asyncIterator]();
while (true) {
const { done, value } = await iterator.next();
if (done) {
return;
}
yield value;
if (signal.aborted) {
return;
}
}
} we can do a lot better, it would be more or less "zero cost" in this case. Generally, in many cases where we use abortableItereator we are using it waiting for small pieces of data, which imo are not worth beeing able to cancel to save a millisecond or what not. Instead, overall, more time is saved if we remove the cancelable promise feature altogether and do the simplest kind of abortable iterator (above). |
For @tuyennhv specific case though I don't think the bottleneck actually is abortable-iterator, because most of the time it is waiting for the next promise to resolve. Judging by the CPU times provided it looks like the noise encryption is what actually is the bottleneck. Perhaps try to optimize this instead? I have made a fork of chainsafe noise encryption where I utilize Node native crypto functions and WASM that gives around 5x improved speed for Node envs @dao-xyz/libp2p-noise |
We discussed this issue in the libp2p triage call today. |
async function* asyncIterableWithAbort<T>(signal: AbortSignal, iterable: AsyncIterable<T>): AsyncGenerator<T, void, undefined> {
const iterator = iterable[Symbol.asyncIterator]();
while (true) {
const { done, value } = await iterator.next();
if (done) {
return;
}
yield value;
if (signal.aborted) {
return;
}
}
} Thinking about this a bit more, the above won't work as it needs the promise returned from Imagine the iterator yields a value, great, we check the abort signal and it's not aborted, then the iterator takes an hour to yield the next value, meanwhile the 'abort' event fires on the abort signal, we won't check the Eg: async function delay (ms) {
return new Promise((resolve) => {
setTimeout(() => resolve(), ms)
})
}
async function * input () {
yield 0
await delay(10)
yield 1
await delay(10)
yield 2
await delay(10)
yield 3
await delay(1000)
yield 4
await delay(10)
yield 5
}
async function * asyncIterableWithAbort(signal, iterable) {
const iterator = iterable[Symbol.asyncIterator]();
while (true) {
const { done, value } = await iterator.next();
if (done) {
return;
}
yield value;
if (signal.aborted) {
return;
}
}
}
const signal = AbortSignal.timeout(100)
const start = Date.now()
for await (const val of asyncIterableWithAbort(signal, input())) {
console.info(val)
}
console.info('took', Date.now() - start, 'ms') % node test.js
0
1
2
3
4
took 1041 ms So it runs for 1041ms even though the timeout was 100ms. Returning a new async generator and calling async function * asyncIterableWithAbort(signal, iterable) {
async function * output () {
yield * iterable
}
const out = output()
const onAbort = () => {
out.return()
// or better:
// out.throw(new Error('Aborted!'))
}
signal.addEventListener('abort', onAbort)
try {
yield * out
} finally {
signal.removeEventListener('abort', onAbort)
}
} % node test.js
0
1
2
3
4
took 1040 ms I wonder if we just want a simpler pattern. E.g. instead of: import { abortableDuplex } from 'abortable-iterator'
async someStreamHandler (data: IncomingStreamData): Promise<void> {
const signal = AbortSignal.timeout(timeout)
const stream = abortableDuplex(data.stream, signal)
await pipe(
stream,
(source) => {
// do some work that we don't want to take forever
}
// ... more code here
} something like: async someStreamHandler (data: IncomingStreamData): Promise<void> {
const signal = AbortSignal.timeout(timeout)
signal.addEventListener('abort', () => {
data.stream.abort(new Error('Timeout!'))
})
await pipe(
data.stream,
(source) => {
// do some work that we don't want to take forever
}
// ... more code here
} |
I think you/we are talking about the same thing here. For promises that are expected to "do some work that we don't want to take forever" you can get away with not having cancellable promises, and can do a naive approach. And for cases where promises are not expected to resolve quickly and you need to be able to cancel them you might need a more robust solution. I did a simple test with the @libp2p/mplex lib where I did a "fast" abortable source approach and only got a 5-10% improvement in the benchmark But to return to @tuyennhv problem, I don't think the abortable-iterator is the problem, but the noise encryption that is creating an iterator that has a lot of "waiting", and for that I suggested a solution that might work quite well and might reduce the wait time with around 80%. |
@marcus-pousette are you interested in taking this on? Note: |
I am kind of working on this problem continuously already, but as stated above I am pretty sure that the bottleneck is not the javascript Promise.race/aborting, so I would argue the premise for this issue is wrong. I did try running my own project where I swapped out all abortable promises to a simpler one and the gains where not significant at all. If you look at the profiling provided, the process that itself is most likely taking up most time is noise encryption. To that I suggested a solution @wemeetagain and Lodestar could try out to see how it improves the CPU-times. In summary, for me (and perhaps also for Lodestar) I would say areas you can improve are
|
Will do. We're currently using
👀 |
Closing this given @marcus-pousette suggestion that these performance bottlenecks are related to the crypto tooling as opposed to I see that @wemeetagain has been testing this on ChainSafe/lodestar#5900 (comment) I have created a new issue to address the discussion around refactoring abortable source usage. |
Version:
0.39.2
Platform:
Linux 5.15.0-25-generic Update libp2p-spdy to version 0.9.0 🚀 #25-Ubuntu SMP Wed Mar 30 15:54:22 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
Subsystem:
Severity:
High
Description:
The
abortableSource()
api is used across libp2p stacks and it takes more than 7.5% of lodestar cpu time, which causes the I/O lag issue in lodestar. I think this is too much considering this is not a core feature when running a node in lodestar.I suggest we need to find a more efficient
abortable
pattern and apply it everywherecc @wemeetagain @mpetrunic @dapplion
Steps to reproduce the error:
Run lodestar, monitor metrics and take a profile from there
The text was updated successfully, but these errors were encountered: