allocation_awareness.asciidoc 4.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104
  1. [[shard-allocation-awareness]]
  2. ==== Shard allocation awareness
  3. You can use custom node attributes as _awareness attributes_ to enable {es}
  4. to take your physical hardware configuration into account when allocating shards.
  5. If {es} knows which nodes are on the same physical server, in the same rack, or
  6. in the same zone, it can distribute the primary shard and its replica shards to
  7. minimise the risk of losing all shard copies in the event of a failure.
  8. When shard allocation awareness is enabled with the
  9. `cluster.routing.allocation.awareness.attributes` setting, shards are only
  10. allocated to nodes that have values set for the specified awareness
  11. attributes. If you use multiple awareness attributes, {es} considers
  12. each attribute separately when allocating shards.
  13. The allocation awareness settings can be configured in
  14. `elasticsearch.yml` and updated dynamically with the
  15. <<cluster-update-settings,cluster-update-settings>> API.
  16. NOTE: The number of attribute values determines how many shard copies are
  17. allocated in each location. If the number of nodes in each location is
  18. unbalanced and there are a lot of replicas, replica shards might be left
  19. unassigned.
  20. [[enabling-awareness]]
  21. ===== Enabling shard allocation awareness
  22. To enable shard allocation awareness:
  23. . Specify the location of each node with a custom node attribute. For example,
  24. if you want Elasticsearch to distribute shards across different racks, you might
  25. set an awareness attribute called `rack_id` in each node's `elasticsearch.yml`
  26. config file.
  27. +
  28. [source,yaml]
  29. --------------------------------------------------------
  30. node.attr.rack_id: rack_one
  31. --------------------------------------------------------
  32. +
  33. You can also set custom attributes when you start a node:
  34. +
  35. [source,sh]
  36. --------------------------------------------------------
  37. ./bin/elasticsearch -Enode.attr.rack_id=rack_one
  38. --------------------------------------------------------
  39. . Tell {es} to take one or more awareness attributes into account when
  40. allocating shards by setting
  41. `cluster.routing.allocation.awareness.attributes` in *every* master-eligible
  42. node's `elasticsearch.yml` config file.
  43. +
  44. --
  45. [source,yaml]
  46. --------------------------------------------------------
  47. cluster.routing.allocation.awareness.attributes: rack_id <1>
  48. --------------------------------------------------------
  49. <1> Specify multiple attributes as a comma-separated list.
  50. --
  51. +
  52. You can also use the
  53. <<cluster-update-settings,cluster-update-settings>> API to set or update
  54. a cluster's awareness attributes.
  55. With this example configuration, if you start two nodes with
  56. `node.attr.rack_id` set to `rack_one` and create an index with 5 primary
  57. shards and 1 replica of each primary, all primaries and replicas are
  58. allocated across the two nodes.
  59. If you add two nodes with `node.attr.rack_id` set to `rack_two`,
  60. {es} moves shards to the new nodes, ensuring (if possible)
  61. that no two copies of the same shard are in the same rack.
  62. If `rack_two` fails and takes down both its nodes, by default {es}
  63. allocates the lost shard copies to nodes in `rack_one`. To prevent multiple
  64. copies of a particular shard from being allocated in the same location, you can
  65. enable forced awareness.
  66. [[forced-awareness]]
  67. ===== Forced awareness
  68. By default, if one location fails, Elasticsearch assigns all of the missing
  69. replica shards to the remaining locations. While you might have sufficient
  70. resources across all locations to host your primary and replica shards, a single
  71. location might be unable to host *ALL* of the shards.
  72. To prevent a single location from being overloaded in the event of a failure,
  73. you can set `cluster.routing.allocation.awareness.force` so no replicas are
  74. allocated until nodes are available in another location.
  75. For example, if you have an awareness attribute called `zone` and configure nodes
  76. in `zone1` and `zone2`, you can use forced awareness to prevent Elasticsearch
  77. from allocating replicas if only one zone is available:
  78. [source,yaml]
  79. -------------------------------------------------------------------
  80. cluster.routing.allocation.awareness.attributes: zone
  81. cluster.routing.allocation.awareness.force.zone.values: zone1,zone2 <1>
  82. -------------------------------------------------------------------
  83. <1> Specify all possible values for the awareness attribute.
  84. With this example configuration, if you start two nodes with `node.attr.zone` set
  85. to `zone1` and create an index with 5 shards and 1 replica, Elasticsearch creates
  86. the index and allocates the 5 primary shards but no replicas. Replicas are
  87. only allocated once nodes with `node.attr.zone` set to `zone2` are available.