JanusGraph become unresponsive after ~10 days #2462
Unanswered
BrunoBerisso
asked this question in
Q&A
Replies: 1 comment 2 replies
-
Do you have any thread dump and/or heap dump before/when OOM happens? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all.
Important: I'm running a modified version of JanusGraph 4.0 that includes the changes described here #2148
I have a three-node JanusGraph cluster backed by another three-node Scylla cluster. We are using
ConfiguredGraphFactory
to create and drop aprox 10 graphs per day and we keep around 70 graphs with historic data, the graphs are read-only and they have a couple of simple indices. The graphs are small with under 1M vertices or edges.The problem is that after around two weeks of running smoothly, one of the nodes start timing out 95% of the requests. It shows high memory usage and the logs showed two types of exceptions that I believe are related:
org.janusgraph.core.JanusGraphException: Could not execute operation due to backend exception
java.lang.OutOfMemoryError: unable to create new native thread
I'm not sure how these two exceptions are related but it seems to be some kind of leak during the drop process that's causing the second exception.
Could not execute operation due to backend exception
Sometimes when a graph is dropped I have this exception in the logs in the node that performed the
JanusGraphFactory.drop(graph);
call. The other nodes don't show any particular activity:java.lang.OutOfMemoryError
This exception appears in the logs after one of the drop() operation is performed and then continue to appear 261 times in a 3min time window with almost no other message interlieved. This is the log entry:
Beta Was this translation helpful? Give feedback.
All reactions