|
@@ -188,20 +188,43 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
|
|
|
========
|
|
|
+`roberta`::::
|
|
|
+(Optional, object)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
|
|
|
++
|
|
|
+.Properties of roberta
|
|
|
+[%collapsible%open]
|
|
|
+========
|
|
|
+`add_prefix_space`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-add-prefix-space]
|
|
|
+
|
|
|
+`max_sequence_length`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
+
|
|
|
+`truncate`::::
|
|
|
+(Optional, string)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
+
|
|
|
+`with_special_tokens`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-with-special-tokens]
|
|
|
+========
|
|
|
`mpnet`::::
|
|
|
(Optional, object)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-mpnet]
|
|
@@ -211,15 +234,15 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
@@ -269,20 +292,43 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
|
|
|
========
|
|
|
+`roberta`::::
|
|
|
+(Optional, object)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
|
|
|
++
|
|
|
+.Properties of roberta
|
|
|
+[%collapsible%open]
|
|
|
+========
|
|
|
+`add_prefix_space`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-add-prefix-space]
|
|
|
+
|
|
|
+`max_sequence_length`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
+
|
|
|
+`truncate`::::
|
|
|
+(Optional, string)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
+
|
|
|
+`with_special_tokens`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-with-special-tokens]
|
|
|
+========
|
|
|
`mpnet`::::
|
|
|
(Optional, object)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-mpnet]
|
|
@@ -292,15 +338,15 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
@@ -343,20 +389,43 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
|
|
|
========
|
|
|
+`roberta`::::
|
|
|
+(Optional, object)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
|
|
|
++
|
|
|
+.Properties of roberta
|
|
|
+[%collapsible%open]
|
|
|
+========
|
|
|
+`add_prefix_space`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-add-prefix-space]
|
|
|
+
|
|
|
+`max_sequence_length`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
+
|
|
|
+`truncate`::::
|
|
|
+(Optional, string)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
+
|
|
|
+`with_special_tokens`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-with-special-tokens]
|
|
|
+========
|
|
|
`mpnet`::::
|
|
|
(Optional, object)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-mpnet]
|
|
@@ -366,15 +435,15 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
@@ -440,24 +509,51 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`span`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-span]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
|
|
|
========
|
|
|
+`roberta`::::
|
|
|
+(Optional, object)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
|
|
|
++
|
|
|
+.Properties of roberta
|
|
|
+[%collapsible%open]
|
|
|
+========
|
|
|
+`add_prefix_space`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-add-prefix-space]
|
|
|
+
|
|
|
+`max_sequence_length`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
+
|
|
|
+`span`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]
|
|
|
+
|
|
|
+`truncate`::::
|
|
|
+(Optional, string)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
+
|
|
|
+`with_special_tokens`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-with-special-tokens]
|
|
|
+========
|
|
|
`mpnet`::::
|
|
|
(Optional, object)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-mpnet]
|
|
@@ -467,19 +563,19 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`span`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-span]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
@@ -522,20 +618,43 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
|
|
|
========
|
|
|
+`roberta`::::
|
|
|
+(Optional, object)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
|
|
|
++
|
|
|
+.Properties of roberta
|
|
|
+[%collapsible%open]
|
|
|
+========
|
|
|
+`add_prefix_space`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-add-prefix-space]
|
|
|
+
|
|
|
+`max_sequence_length`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
+
|
|
|
+`truncate`::::
|
|
|
+(Optional, string)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
+
|
|
|
+`with_special_tokens`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-with-special-tokens]
|
|
|
+========
|
|
|
`mpnet`::::
|
|
|
(Optional, object)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-mpnet]
|
|
@@ -545,15 +664,15 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
@@ -611,20 +730,43 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
|
|
|
========
|
|
|
+`roberta`::::
|
|
|
+(Optional, object)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
|
|
|
++
|
|
|
+.Properties of roberta
|
|
|
+[%collapsible%open]
|
|
|
+========
|
|
|
+`add_prefix_space`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-add-prefix-space]
|
|
|
+
|
|
|
+`max_sequence_length`::::
|
|
|
+(Optional, integer)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
+
|
|
|
+`truncate`::::
|
|
|
+(Optional, string)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
+
|
|
|
+`with_special_tokens`::::
|
|
|
+(Optional, boolean)
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta-with-special-tokens]
|
|
|
+========
|
|
|
`mpnet`::::
|
|
|
(Optional, object)
|
|
|
include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-mpnet]
|
|
@@ -634,15 +776,15 @@ include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizati
|
|
|
========
|
|
|
`do_lower_case`::::
|
|
|
(Optional, boolean)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-do-lower-case]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-do-lower-case]
|
|
|
|
|
|
`max_sequence_length`::::
|
|
|
(Optional, integer)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-max-sequence-length]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-max-sequence-length]
|
|
|
|
|
|
`truncate`::::
|
|
|
(Optional, string)
|
|
|
-include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-truncate]
|
|
|
+include::{es-repo-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
|
|
|
|
|
|
`with_special_tokens`::::
|
|
|
(Optional, boolean)
|