| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114 | [[allocation-awareness]]=== Shard Allocation AwarenessWhen running nodes on multiple VMs on the same physical server, on multipleracks, or across multiple awareness zones, it is more likely that two nodes onthe same physical server, in the same rack, or in the same awareness zone willcrash at the same time, rather than two unrelated nodes crashingsimultaneously.If Elasticsearch is _aware_ of the physical configuration of your hardware, itcan ensure that the primary shard and its replica shards are spread acrossdifferent physical servers, racks, or zones, to minimise the risk of losingall shard copies at the same time.The shard allocation awareness settings allow you to tell Elasticsearch aboutyour hardware configuration.As an example, let's assume we have several racks.  When we start a node, wecan tell it which rack it is in by assigning it an arbitrary metadataattribute called `rack_id` -- we could use any attribute name.  For example:[source,sh]----------------------./bin/elasticsearch -Enode.attr.rack_id=rack_one <1>----------------------<1> This setting could also be specified in the `elasticsearch.yml` config file.Now, we need to setup _shard allocation awareness_  by telling Elasticsearchwhich attributes to use.  This can be configured in the `elasticsearch.yml`file on *all* master-eligible nodes, or it can be set (and changed) with the<<cluster-update-settings,cluster-update-settings>> API.For our example, we'll set the value in the config file:[source,yaml]--------------------------------------------------------cluster.routing.allocation.awareness.attributes: rack_id--------------------------------------------------------With this config in place, let's say we start two nodes with `node.attr.rack_id`set to `rack_one`, and we create an index with 5 primary shards and 1 replicaof each primary.  All primaries and replicas are allocated across the twonodes.Now, if we start two more nodes with `node.attr.rack_id` set to `rack_two`,Elasticsearch will move shards across to the new nodes, ensuring (if possible)that no two copies of the same shard will be in the same rack. However if `rack_two`were to fail, taking down both of its nodes, Elasticsearch will still allocate the lostshard copies to nodes in `rack_one`. .Prefer local shards*********************************************When executing search or GET requests, with shard awareness enabled,Elasticsearch will prefer using local shards -- shards in the same awarenessgroup -- to execute the request. This is usually faster than crossing racks orawareness zones.*********************************************Multiple awareness attributes can be specified, in which case the combinationof values from each attribute is considered to be a separate value.[source,yaml]-------------------------------------------------------------cluster.routing.allocation.awareness.attributes: rack_id,zone-------------------------------------------------------------NOTE: When using awareness attributes, shards will not be allocated tonodes that don't have values set for those attributes.NOTE: Number of primary/replica of a shard allocated on a specific groupof nodes with the same awareness attribute value is determined by the numberof attribute values. When the number of nodes in groups is unbalanced andthere are many replicas, replica shards may be left unassigned.[float][[forced-awareness]]=== Forced AwarenessImagine that you have two awareness zones and enough hardware across the twozones to host all of your primary and replica shards.  But perhaps thehardware in a single zone, while sufficient to host half the shards, would beunable to host *ALL* the shards.With ordinary awareness, if one zone lost contact with the other zone,Elasticsearch would assign all of the missing replica shards to a single zone.But in this example, this sudden extra load would cause the hardware in theremaining zone to be overloaded.Forced awareness solves this problem by *NEVER* allowing copies of the sameshard to be allocated to the same zone.For example, lets say we have an awareness attribute called `zone`, andwe know we are going to have two zones, `zone1` and `zone2`. Here is howwe can force awareness on a node:[source,yaml]-------------------------------------------------------------------cluster.routing.allocation.awareness.force.zone.values: zone1,zone2 <1>cluster.routing.allocation.awareness.attributes: zone-------------------------------------------------------------------<1> We must list all possible values that the `zone` attribute can have.Now, if we start 2 nodes with `node.attr.zone` set to `zone1` and create an indexwith 5 shards and 1 replica. The index will be created, but only the 5 primaryshards will be allocated (with no replicas). Only when we start more shardswith `node.attr.zone` set to `zone2` will the replicas be allocated.The `cluster.routing.allocation.awareness.*` settings can all be updateddynamically on a live cluster with the<<cluster-update-settings,cluster-update-settings>> API.
 |