network-timeouts.asciidoc 3.0 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253
  1. tag::troubleshooting-network-timeouts-gc-vm[]
  2. * GC pauses are recorded in the GC logs that {es} emits by default, and also
  3. usually by the `JvmMonitorService` in the main node logs. Use these logs to
  4. confirm whether or not the node is experiencing high heap usage with long GC
  5. pauses. If so, <<high-jvm-memory-pressure,the troubleshooting guide for high
  6. heap usage>> has some suggestions for further investigation but typically you
  7. will need to capture a heap dump during a time of high heap usage to fully
  8. understand the problem.
  9. * VM pauses also affect other processes on the same host. A VM pause also
  10. typically causes a discontinuity in the system clock, which {es} will report in
  11. its logs. If you see evidence of other processes pausing at the same time, or
  12. unexpected clock discontinuities, investigate the infrastructure on which you
  13. are running {es}.
  14. end::troubleshooting-network-timeouts-gc-vm[]
  15. tag::troubleshooting-network-timeouts-packet-capture-elections[]
  16. * Packet captures will reveal system-level and network-level faults, especially
  17. if you capture the network traffic simultaneously at all relevant nodes. You
  18. should be able to observe any retransmissions, packet loss, or other delays on
  19. the connections between the nodes.
  20. end::troubleshooting-network-timeouts-packet-capture-elections[]
  21. tag::troubleshooting-network-timeouts-packet-capture-fault-detection[]
  22. * Packet captures will reveal system-level and network-level faults, especially
  23. if you capture the network traffic simultaneously at the elected master and the
  24. faulty node. The connection used for follower checks is not used for any other
  25. traffic so it can be easily identified from the flow pattern alone, even if TLS
  26. is in use: almost exactly every second there will be a few hundred bytes sent
  27. each way, first the request by the master and then the response by the
  28. follower. You should be able to observe any retransmissions, packet loss, or
  29. other delays on such a connection.
  30. end::troubleshooting-network-timeouts-packet-capture-fault-detection[]
  31. tag::troubleshooting-network-timeouts-threads[]
  32. * Long waits for particular threads to be available can be identified by taking
  33. stack dumps (for example, using `jstack`) or a profiling trace (for example,
  34. using Java Flight Recorder) in the few seconds leading up to the relevant log
  35. message.
  36. +
  37. The <<cluster-nodes-hot-threads>> API sometimes yields useful information, but
  38. bear in mind that this API also requires a number of `transport_worker` and
  39. `generic` threads across all the nodes in the cluster. The API may be affected
  40. by the very problem you're trying to diagnose. `jstack` is much more reliable
  41. since it doesn't require any JVM threads.
  42. +
  43. The threads involved in discovery and cluster membership are mainly
  44. `transport_worker` and `cluster_coordination` threads, for which there should
  45. never be a long wait. There may also be evidence of long waits for threads in
  46. the {es} logs, particularly looking at warning logs from
  47. `org.elasticsearch.transport.InboundHandler`. See
  48. <<modules-network-threading-model>> for more information.
  49. end::troubleshooting-network-timeouts-threads[]