disk_allocator.asciidoc 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103
  1. [[disk-allocator]]
  2. === Disk-based shard allocation
  3. Elasticsearch considers the available disk space on a node before deciding
  4. whether to allocate new shards to that node or to actively relocate shards away
  5. from that node.
  6. Below are the settings that can be configured in the `elasticsearch.yml` config
  7. file or updated dynamically on a live cluster with the
  8. <<cluster-update-settings,cluster-update-settings>> API:
  9. `cluster.routing.allocation.disk.threshold_enabled`::
  10. Defaults to `true`. Set to `false` to disable the disk allocation decider.
  11. `cluster.routing.allocation.disk.watermark.low`::
  12. Controls the low watermark for disk usage. It defaults to `85%`, meaning
  13. that Elasticsearch will not allocate shards to nodes that have more than
  14. 85% disk used. It can also be set to an absolute byte value (like `500mb`)
  15. to prevent Elasticsearch from allocating shards if less than the specified
  16. amount of space is available. This setting has no effect on the primary
  17. shards of newly-created indices or, specifically, any shards that have
  18. never previously been allocated.
  19. `cluster.routing.allocation.disk.watermark.high`::
  20. Controls the high watermark. It defaults to `90%`, meaning that
  21. Elasticsearch will attempt to relocate shards away from a node whose disk
  22. usage is above 90%. It can also be set to an absolute byte value (similarly
  23. to the low watermark) to relocate shards away from a node if it has less
  24. than the specified amount of free space. This setting affects the
  25. allocation of all shards, whether previously allocated or not.
  26. `cluster.routing.allocation.disk.watermark.flood_stage`::
  27. +
  28. --
  29. Controls the flood stage watermark. It defaults to 95%, meaning that
  30. Elasticsearch enforces a read-only index block
  31. (`index.blocks.read_only_allow_delete`) on every index that has one or more
  32. shards allocated on the node that has at least one disk exceeding the flood
  33. stage. This is a last resort to prevent nodes from running out of disk space.
  34. The index block is automatically released once the disk utilization falls below
  35. the high watermark.
  36. NOTE: You can not mix the usage of percentage values and byte values within
  37. these settings. Either all are set to percentage values, or all are set to byte
  38. values. This is so that we can we validate that the settings are internally
  39. consistent (that is, the low disk threshold is not more than the high disk
  40. threshold, and the high disk threshold is not more than the flood stage
  41. threshold).
  42. An example of resetting the read-only index block on the `twitter` index:
  43. [source,console]
  44. --------------------------------------------------
  45. PUT /twitter/_settings
  46. {
  47. "index.blocks.read_only_allow_delete": null
  48. }
  49. --------------------------------------------------
  50. // TEST[setup:twitter]
  51. --
  52. `cluster.info.update.interval`::
  53. How often Elasticsearch should check on disk usage for each node in the
  54. cluster. Defaults to `30s`.
  55. NOTE: Percentage values refer to used disk space, while byte values refer to
  56. free disk space. This can be confusing, since it flips the meaning of high and
  57. low. For example, it makes sense to set the low watermark to 10gb and the high
  58. watermark to 5gb, but not the other way around.
  59. An example of updating the low watermark to at least 100 gigabytes free, a high
  60. watermark of at least 50 gigabytes free, and a flood stage watermark of 10
  61. gigabytes free, and updating the information about the cluster every minute:
  62. [source,console]
  63. --------------------------------------------------
  64. PUT _cluster/settings
  65. {
  66. "transient": {
  67. "cluster.routing.allocation.disk.watermark.low": "100gb",
  68. "cluster.routing.allocation.disk.watermark.high": "50gb",
  69. "cluster.routing.allocation.disk.watermark.flood_stage": "10gb",
  70. "cluster.info.update.interval": "1m"
  71. }
  72. }
  73. --------------------------------------------------
  74. {es} accounts for the future disk usage of ongoing shard relocations and
  75. recoveries to help prevent these shard movements from breaching a watermark.
  76. This mechanism may double-count some data that has already been relocated onto
  77. a node. For instance, if a relocation of a 100GB shard is 90% complete then
  78. {es} has copied 90GB of data onto the target node. This 90GB consumes disk
  79. space and will be reflected in the node's disk usage statistics. However {es}
  80. also treats the relocation as if it will consume another full 100GB in the
  81. future, even though the shard may really only consume a further 10GB of space.
  82. If the node's disks are close to a watermark then this may temporarily prevent
  83. other shards from moving onto the same node. Eventually the relocation will
  84. complete and then {es} will use the node's true disk usage statistics again.