misc.asciidoc 5.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150
  1. [[misc-cluster]]
  2. === Miscellaneous cluster settings
  3. [[cluster-read-only]]
  4. ==== Metadata
  5. An entire cluster may be set to read-only with the following _dynamic_ setting:
  6. `cluster.blocks.read_only`::
  7. Make the whole cluster read only (indices do not accept write
  8. operations), metadata is not allowed to be modified (create or delete
  9. indices).
  10. `cluster.blocks.read_only_allow_delete`::
  11. Identical to `cluster.blocks.read_only` but allows to delete indices
  12. to free up resources.
  13. WARNING: Don't rely on this setting to prevent changes to your cluster. Any
  14. user with access to the <<cluster-update-settings,cluster-update-settings>>
  15. API can make the cluster read-write again.
  16. [[cluster-shard-limit]]
  17. ==== Cluster Shard Limit
  18. There is a soft limit on the number of shards in a cluster, based on the number
  19. of nodes in the cluster. This is intended to prevent operations which may
  20. unintentionally destabilize the cluster.
  21. If an operation, such as creating a new index, restoring a snapshot of an index,
  22. or opening a closed index would lead to the number of shards in the cluster
  23. going over this limit, the operation will fail with an error indicating the
  24. shard limit.
  25. If the cluster is already over the limit, due to changes in node membership or
  26. setting changes, all operations that create or open indices will fail until
  27. either the limit is increased as described below, or some indices are
  28. <<indices-open-close,closed>> or <<indices-delete-index,deleted>> to bring the
  29. number of shards below the limit.
  30. Replicas count towards this limit, but closed indexes do not. An index with 5
  31. primary shards and 2 replicas will be counted as 15 shards. Any closed index
  32. is counted as 0, no matter how many shards and replicas it contains.
  33. The limit defaults to 1,000 shards per data node, and can be dynamically
  34. adjusted using the following property:
  35. `cluster.max_shards_per_node`::
  36. Controls the number of shards allowed in the cluster per data node.
  37. For example, a 3-node cluster with the default setting would allow 3,000 shards
  38. total, across all open indexes. If the above setting is changed to 1,500, then
  39. the cluster would allow 4,500 shards total.
  40. NOTE: If there are no data nodes in the cluster, the limit will not be enforced.
  41. This allows the creation of indices during cluster creation if dedicated master
  42. nodes are set up before data nodes.
  43. [[user-defined-data]]
  44. ==== User Defined Cluster Metadata
  45. User-defined metadata can be stored and retrieved using the Cluster Settings API.
  46. This can be used to store arbitrary, infrequently-changing data about the cluster
  47. without the need to create an index to store it. This data may be stored using
  48. any key prefixed with `cluster.metadata.`. For example, to store the email
  49. address of the administrator of a cluster under the key `cluster.metadata.administrator`,
  50. issue this request:
  51. [source,js]
  52. -------------------------------
  53. PUT /_cluster/settings
  54. {
  55. "persistent": {
  56. "cluster.metadata.administrator": "sysadmin@example.com"
  57. }
  58. }
  59. -------------------------------
  60. // CONSOLE
  61. IMPORTANT: User-defined cluster metadata is not intended to store sensitive or
  62. confidential information. Any information stored in user-defined cluster
  63. metadata will be viewable by anyone with access to the
  64. <<cluster-get-settings,Cluster Get Settings>> API, and is recorded in the
  65. {es} logs.
  66. [[cluster-max-tombstones]]
  67. ==== Index Tombstones
  68. The cluster state maintains index tombstones to explicitly denote indices that
  69. have been deleted. The number of tombstones maintained in the cluster state is
  70. controlled by the following property, which cannot be updated dynamically:
  71. `cluster.indices.tombstones.size`::
  72. Index tombstones prevent nodes that are not part of the cluster when a delete
  73. occurs from joining the cluster and reimporting the index as though the delete
  74. was never issued. To keep the cluster state from growing huge we only keep the
  75. last `cluster.indices.tombstones.size` deletes, which defaults to 500. You can
  76. increase it if you expect nodes to be absent from the cluster and miss more
  77. than 500 deletes. We think that is rare, thus the default. Tombstones don't take
  78. up much space, but we also think that a number like 50,000 is probably too big.
  79. [[cluster-logger]]
  80. ==== Logger
  81. The settings which control logging can be updated dynamically with the
  82. `logger.` prefix. For instance, to increase the logging level of the
  83. `indices.recovery` module to `DEBUG`, issue this request:
  84. [source,js]
  85. -------------------------------
  86. PUT /_cluster/settings
  87. {
  88. "transient": {
  89. "logger.org.elasticsearch.indices.recovery": "DEBUG"
  90. }
  91. }
  92. -------------------------------
  93. // CONSOLE
  94. [[persistent-tasks-allocation]]
  95. ==== Persistent Tasks Allocations
  96. Plugins can create a kind of tasks called persistent tasks. Those tasks are
  97. usually long-live tasks and are stored in the cluster state, allowing the
  98. tasks to be revived after a full cluster restart.
  99. Every time a persistent task is created, the master nodes takes care of
  100. assigning the task to a node of the cluster, and the assigned node will then
  101. pick up the task and execute it locally. The process of assigning persistent
  102. tasks to nodes is controlled by the following property, which can be updated
  103. dynamically:
  104. `cluster.persistent_tasks.allocation.enable`::
  105. +
  106. --
  107. Enable or disable allocation for persistent tasks:
  108. * `all` - (default) Allows persistent tasks to be assigned to nodes
  109. * `none` - No allocations are allowed for any type of persistent task
  110. This setting does not affect the persistent tasks that are already being executed.
  111. Only newly created persistent tasks, or tasks that must be reassigned (after a node
  112. left the cluster, for example), are impacted by this setting.
  113. --