-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Possible finagle-chirper crash #106
Comments
cc @Fithos |
I wasn't able to reproduce this error on Linux and MacOS. |
Unfortunately, neither was I on my machine. I am in contact with Tom Rodriguez who reported it, who sees it happening on some cluster machines (apparently it's an issue that happens both on HotSpot and on GraalVM). |
I've really only seen this on a fairly large machine. It's an Oracle x5-2 with 72 cores running Oracle Linux Server release 6.8. i've also seen problems with running out of file descriptors though that was with an older version of this. The limit is 65536 so it would have to be leaking fds pretty badly for this to occur. Anything I can try to help with debugging? |
The benchmark has a leaking file descriptor issue. Example exception dump: Caused by: java.lang.reflect.InvocationTargetException |
Seen more instances of this on regular machines (e.g. measly 8 core 64GB RAM). At the time of the crash, the benchmark has quite a lot of TCP sockets open. Sometimes, this makes the benchmark hang instead of crash, complicating measurements. |
... also see constantly increasing number of threads (right now ~300 threads at iteration 400), if it helps, most thread instances are named UnboundedFuturePool-Something. |
@Fithos Can you confirm the file descriptor issue, and thread growth? |
It's somewhat surprising that the thread count seems to be converging. But, also, it drops at the very end, and the |
Maybe it's because I'm also not seeing when other Clients in Finagle are designed to be long-lived and potentially shared across your application. They operate better and better the longer you keep them around (connections are materialized, a past performance about errors/latencies is recorded, etc). |
Hi Vladimir! Thanks for the pointer! |
That's right, @axel22. Just call |
Some users reported seeing this on some machines:
The text was updated successfully, but these errors were encountered: