Skip to content

Commit 9dc2fef

Browse files
authored
Add a note about using the retry_failed flag before accepting data loss (#29160)
1 parent 2e1e862 commit 9dc2fef

File tree

1 file changed

+20
-14
lines changed

1 file changed

+20
-14
lines changed

docs/reference/cluster/reroute.asciidoc

Lines changed: 20 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -83,8 +83,26 @@ Reasons why a primary shard cannot be automatically allocated include the follow
8383
the cluster. To prevent data loss, the system does not automatically promote a stale
8484
shard copy to primary.
8585

86-
As a manual override, two commands to forcefully allocate primary shards
87-
are available:
86+
[float]
87+
=== Retry failed shards
88+
89+
The cluster will attempt to allocate a shard a maximum of
90+
`index.allocation.max_retries` times in a row (defaults to `5`), before giving
91+
up and leaving the shard unallocated. This scenario can be caused by
92+
structural problems such as having an analyzer which refers to a stopwords
93+
file which doesn't exist on all nodes.
94+
95+
Once the problem has been corrected, allocation can be manually retried by
96+
calling the <<cluster-reroute,`reroute`>> API with `?retry_failed`, which
97+
will attempt a single retry round for these shards.
98+
99+
[float]
100+
=== Forced allocation on unrecoverable errors
101+
102+
The following two commands are dangerous and may result in data loss. They are
103+
meant to be used in cases where the original data can not be recovered and the cluster
104+
administrator accepts the loss. If you have suffered a temporary issue that has been
105+
fixed, please see the `retry_failed` flag described above.
88106

89107
`allocate_stale_primary`::
90108
Allocate a primary shard to a node that holds a stale copy. Accepts the
@@ -108,15 +126,3 @@ are available:
108126
this command requires the special field `accept_data_loss` to be
109127
explicitly set to `true` for it to work.
110128

111-
[float]
112-
=== Retry failed shards
113-
114-
The cluster will attempt to allocate a shard a maximum of
115-
`index.allocation.max_retries` times in a row (defaults to `5`), before giving
116-
up and leaving the shard unallocated. This scenario can be caused by
117-
structural problems such as having an analyzer which refers to a stopwords
118-
file which doesn't exist on all nodes.
119-
120-
Once the problem has been corrected, allocation can be manually retried by
121-
calling the <<cluster-reroute,`reroute`>> API with `?retry_failed`, which
122-
will attempt a single retry round for these shards.

0 commit comments

Comments
 (0)