significanttext-aggregation.asciidoc 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493
  1. [[search-aggregations-bucket-significanttext-aggregation]]
  2. === Significant Text Aggregation
  3. experimental[]
  4. An aggregation that returns interesting or unusual occurrences of free-text terms in a set.
  5. It is like the <<search-aggregations-bucket-significantterms-aggregation,significant terms>> aggregation but differs in that:
  6. * It is specifically designed for use on type `text` fields
  7. * It does not require field data or doc-values
  8. * It re-analyzes text content on-the-fly meaning it can also filter duplicate sections of
  9. noisy text that otherwise tend to skew statistics.
  10. WARNING: Re-analyzing _large_ result sets will require a lot of time and memory. It is recommended that the significant_text
  11. aggregation is used as a child of either the <<search-aggregations-bucket-sampler-aggregation,sampler>> or
  12. <<search-aggregations-bucket-diversified-sampler-aggregation,diversified sampler>> aggregation to limit the analysis
  13. to a _small_ selection of top-matching documents e.g. 200. This will typically improve speed, memory use and quality of
  14. results.
  15. .Example use cases:
  16. * Suggesting "H5N1" when users search for "bird flu" to help expand queries
  17. * Suggesting keywords relating to stock symbol $ATI for use in an automated news classifier
  18. In these cases the words being selected are not simply the most popular terms in results. The most popular words tend to be
  19. very boring (_and, of, the, we, I, they_ ...).
  20. The significant words are the ones that have undergone a significant change in popularity measured between a _foreground_ and _background_ set.
  21. If the term "H5N1" only exists in 5 documents in a 10 million document index and yet is found in 4 of the 100 documents that make up a user's search results
  22. that is significant and probably very relevant to their search. 5/10,000,000 vs 4/100 is a big swing in frequency.
  23. experimental[The `significant_text` aggregation is new and may change in non-backwards compatible ways if we add further text-analysis features e.g. phrase detection]
  24. ==== Basic use
  25. In the typical use case, the _foreground_ set of interest is a selection of the top-matching search results for a query
  26. and the _background_set used for statistical comparisons is the index or indices from which the results were gathered.
  27. Example:
  28. [source,js]
  29. --------------------------------------------------
  30. GET news/_search
  31. {
  32. "query" : {
  33. "match" : {"content" : "Bird flu"}
  34. },
  35. "aggregations" : {
  36. "my_sample" : {
  37. "sampler" : {
  38. "shard_size" : 100
  39. },
  40. "aggregations": {
  41. "keywords" : {
  42. "significant_text" : { "field" : "content" }
  43. }
  44. }
  45. }
  46. }
  47. }
  48. --------------------------------------------------
  49. // CONSOLE
  50. // TEST[setup:news]
  51. Response:
  52. [source,js]
  53. --------------------------------------------------
  54. {
  55. "took": 9,
  56. "timed_out": false,
  57. "_shards": ...,
  58. "hits": ...,
  59. "aggregations" : {
  60. "my_sample": {
  61. "doc_count": 100,
  62. "keywords" : {
  63. "doc_count": 100,
  64. "buckets" : [
  65. {
  66. "key": "h5n1",
  67. "doc_count": 4,
  68. "score": 4.71235374214817,
  69. "bg_count": 5
  70. }
  71. ...
  72. ]
  73. }
  74. }
  75. }
  76. }
  77. --------------------------------------------------
  78. // TESTRESPONSE[skip:historically skipped]
  79. The results show that "h5n1" is one of several terms strongly associated with bird flu.
  80. It only occurs 5 times in our index as a whole (see the `bg_count`) and yet 4 of these
  81. were lucky enough to appear in our 100 document sample of "bird flu" results. That suggests
  82. a significant word and one which the user can potentially add to their search.
  83. ==== Dealing with noisy data using `filter_duplicate_text`
  84. Free-text fields often contain a mix of original content and mechanical copies of text (cut-and-paste biographies, email reply chains,
  85. retweets, boilerplate headers/footers, page navigation menus, sidebar news links, copyright notices, standard disclaimers, addresses).
  86. In real-world data these duplicate sections of text tend to feature heavily in `significant_text` results if they aren't filtered out.
  87. Filtering near-duplicate text is a difficult task at index-time but we can cleanse the data on-the-fly at query time using the
  88. `filter_duplicate_text` setting.
  89. First let's look at an unfiltered real-world example using the http://research.signalmedia.co/newsir16/signal-dataset.html[Signal media dataset] of
  90. a million news articles covering a wide variety of news. Here are the raw significant text results for a search for the articles
  91. mentioning "elasticsearch":
  92. [source,js]
  93. --------------------------------------------------
  94. {
  95. ...
  96. "aggregations": {
  97. "sample": {
  98. "doc_count": 35,
  99. "keywords": {
  100. "doc_count": 35,
  101. "buckets": [
  102. {
  103. "key": "elasticsearch",
  104. "doc_count": 35,
  105. "score": 28570.428571428572,
  106. "bg_count": 35
  107. },
  108. ...
  109. {
  110. "key": "currensee",
  111. "doc_count": 8,
  112. "score": 6530.383673469388,
  113. "bg_count": 8
  114. },
  115. ...
  116. {
  117. "key": "pozmantier",
  118. "doc_count": 4,
  119. "score": 3265.191836734694,
  120. "bg_count": 4
  121. },
  122. ...
  123. }
  124. --------------------------------------------------
  125. // NOTCONSOLE
  126. The uncleansed documents have thrown up some odd-looking terms that are, on the face of it, statistically
  127. correlated with appearances of our search term "elasticsearch" e.g. "pozmantier".
  128. We can drill down into examples of these documents to see why pozmantier is connected using this query:
  129. [source,js]
  130. --------------------------------------------------
  131. GET news/_search
  132. {
  133. "query": {
  134. "simple_query_string": {
  135. "query": "+elasticsearch +pozmantier"
  136. }
  137. },
  138. "_source": [
  139. "title",
  140. "source"
  141. ],
  142. "highlight": {
  143. "fields": {
  144. "content": {}
  145. }
  146. }
  147. }
  148. --------------------------------------------------
  149. // CONSOLE
  150. // TEST[setup:news]
  151. The results show a series of very similar news articles about a judging panel for a number of tech projects:
  152. [source,js]
  153. --------------------------------------------------
  154. {
  155. ...
  156. "hits": {
  157. "hits": [
  158. {
  159. ...
  160. "_source": {
  161. "source": "Presentation Master",
  162. "title": "T.E.N. Announces Nominees for the 2015 ISE® North America Awards"
  163. },
  164. "highlight": {
  165. "content": [
  166. "City of San Diego Mike <em>Pozmantier</em>, Program Manager, Cyber Security Division, Department of",
  167. " Janus, Janus <em>ElasticSearch</em> Security Visualization Engine "
  168. ]
  169. }
  170. },
  171. {
  172. ...
  173. "_source": {
  174. "source": "RCL Advisors",
  175. "title": "T.E.N. Announces Nominees for the 2015 ISE(R) North America Awards"
  176. },
  177. "highlight": {
  178. "content": [
  179. "Mike <em>Pozmantier</em>, Program Manager, Cyber Security Division, Department of Homeland Security S&T",
  180. "Janus, Janus <em>ElasticSearch</em> Security Visualization Engine"
  181. ]
  182. }
  183. },
  184. ...
  185. --------------------------------------------------
  186. // NOTCONSOLE
  187. Mike Pozmantier was one of many judges on a panel and elasticsearch was used in one of many projects being judged.
  188. As is typical, this lengthy press release was cut-and-paste by a variety of news sites and consequently any rare names, numbers or
  189. typos they contain become statistically correlated with our matching query.
  190. Fortunately similar documents tend to rank similarly so as part of examining the stream of top-matching documents the significant_text
  191. aggregation can apply a filter to remove sequences of any 6 or more tokens that have already been seen. Let's try this same query now but
  192. with the `filter_duplicate_text` setting turned on:
  193. [source,js]
  194. --------------------------------------------------
  195. GET news/_search
  196. {
  197. "query": {
  198. "match": {
  199. "content": "elasticsearch"
  200. }
  201. },
  202. "aggs": {
  203. "sample": {
  204. "sampler": {
  205. "shard_size": 100
  206. },
  207. "aggs": {
  208. "keywords": {
  209. "significant_text": {
  210. "field": "content",
  211. "filter_duplicate_text": true
  212. }
  213. }
  214. }
  215. }
  216. }
  217. }
  218. --------------------------------------------------
  219. // CONSOLE
  220. // TEST[setup:news]
  221. The results from analysing our deduplicated text are obviously of higher quality to anyone familiar with the elastic stack:
  222. [source,js]
  223. --------------------------------------------------
  224. {
  225. ...
  226. "aggregations": {
  227. "sample": {
  228. "doc_count": 35,
  229. "keywords": {
  230. "doc_count": 35,
  231. "buckets": [
  232. {
  233. "key": "elasticsearch",
  234. "doc_count": 22,
  235. "score": 11288.001166180758,
  236. "bg_count": 35
  237. },
  238. {
  239. "key": "logstash",
  240. "doc_count": 3,
  241. "score": 1836.648979591837,
  242. "bg_count": 4
  243. },
  244. {
  245. "key": "kibana",
  246. "doc_count": 3,
  247. "score": 1469.3020408163263,
  248. "bg_count": 5
  249. }
  250. ]
  251. }
  252. }
  253. }
  254. }
  255. --------------------------------------------------
  256. // NOTCONSOLE
  257. Mr Pozmantier and other one-off associations with elasticsearch no longer appear in the aggregation
  258. results as a consequence of copy-and-paste operations or other forms of mechanical repetition.
  259. If your duplicate or near-duplicate content is identifiable via a single-value indexed field (perhaps
  260. a hash of the article's `title` text or an `original_press_release_url` field) then it would be more
  261. efficient to use a parent <<search-aggregations-bucket-diversified-sampler-aggregation,diversified sampler>> aggregation
  262. to eliminate these documents from the sample set based on that single key. The less duplicate content you can feed into
  263. the significant_text aggregation up front the better in terms of performance.
  264. .How are the significance scores calculated?
  265. **********************************
  266. The numbers returned for scores are primarily intended for ranking different suggestions sensibly rather than something easily
  267. understood by end users. The scores are derived from the doc frequencies in _foreground_ and _background_ sets. In brief, a
  268. term is considered significant if there is a noticeable difference in the frequency in which a term appears in the subset and
  269. in the background. The way the terms are ranked can be configured, see "Parameters" section.
  270. **********************************
  271. .Use the _"like this but not this"_ pattern
  272. **********************************
  273. You can spot mis-categorized content by first searching a structured field e.g. `category:adultMovie` and use significant_text on the
  274. text "movie_description" field. Take the suggested words (I'll leave them to your imagination) and then search for all movies NOT marked as category:adultMovie but containing these keywords.
  275. You now have a ranked list of badly-categorized movies that you should reclassify or at least remove from the "familyFriendly" category.
  276. The significance score from each term can also provide a useful `boost` setting to sort matches.
  277. Using the `minimum_should_match` setting of the `terms` query with the keywords will help control the balance of precision/recall in the result set i.e
  278. a high setting would have a small number of relevant results packed full of keywords and a setting of "1" would produce a more exhaustive results set with all documents containing _any_ keyword.
  279. **********************************
  280. ==== Limitations
  281. ===== No support for child aggregations
  282. The significant_text aggregation intentionally does not support the addition of child aggregations because:
  283. * It would come with a high memory cost
  284. * It isn't a generally useful feature and there is a workaround for those that need it
  285. The volume of candidate terms is generally very high and these are pruned heavily before the final
  286. results are returned. Supporting child aggregations would generate additional churn and be inefficient.
  287. Clients can always take the heavily-trimmed set of results from a `significant_text` request and
  288. make a subsequent follow-up query using a `terms` aggregation with an `include` clause and child
  289. aggregations to perform further analysis of selected keywords in a more efficient fashion.
  290. ===== No support for nested objects
  291. The significant_text aggregation currently also cannot be used with text fields in
  292. nested objects, because it works with the document JSON source. This makes this
  293. feature inefficient when matching nested docs from stored JSON given a matching
  294. Lucene docID.
  295. ===== Approximate counts
  296. The counts of how many documents contain a term provided in results are based on summing the samples returned from each shard and
  297. as such may be:
  298. * low if certain shards did not provide figures for a given term in their top sample
  299. * high when considering the background frequency as it may count occurrences found in deleted documents
  300. Like most design decisions, this is the basis of a trade-off in which we have chosen to provide fast performance at the cost of some (typically small) inaccuracies.
  301. However, the `size` and `shard size` settings covered in the next section provide tools to help control the accuracy levels.
  302. ==== Parameters
  303. ===== Significance heuristics
  304. This aggregation supports the same scoring heuristics (JLH, mutual_information, gnd, chi_square etc) as the <<search-aggregations-bucket-significantterms-aggregation,significant terms>> aggregation
  305. ===== Size & Shard Size
  306. The `size` parameter can be set to define how many term buckets should be returned out of the overall terms list. By
  307. default, the node coordinating the search process will request each shard to provide its own top term buckets
  308. and once all shards respond, it will reduce the results to the final list that will then be returned to the client.
  309. If the number of unique terms is greater than `size`, the returned list can be slightly off and not accurate
  310. (it could be that the term counts are slightly off and it could even be that a term that should have been in the top
  311. size buckets was not returned).
  312. To ensure better accuracy a multiple of the final `size` is used as the number of terms to request from each shard
  313. using a heuristic based on the number of shards. To take manual control of this setting the `shard_size` parameter
  314. can be used to control the volumes of candidate terms produced by each shard.
  315. Low-frequency terms can turn out to be the most interesting ones once all results are combined so the
  316. significant_terms aggregation can produce higher-quality results when the `shard_size` parameter is set to
  317. values significantly higher than the `size` setting. This ensures that a bigger volume of promising candidate terms are given
  318. a consolidated review by the reducing node before the final selection. Obviously large candidate term lists
  319. will cause extra network traffic and RAM usage so this is quality/cost trade off that needs to be balanced. If `shard_size` is set to -1 (the default) then `shard_size` will be automatically estimated based on the number of shards and the `size` parameter.
  320. NOTE: `shard_size` cannot be smaller than `size` (as it doesn't make much sense). When it is, elasticsearch will
  321. override it and reset it to be equal to `size`.
  322. ===== Minimum document count
  323. It is possible to only return terms that match more than a configured number of hits using the `min_doc_count` option.
  324. The Default value is 3.
  325. Terms that score highly will be collected on a shard level and merged with the terms collected from other shards in a second step.
  326. However, the shard does not have the information about the global term frequencies available. The decision if a term is added to a
  327. candidate list depends only on the score computed on the shard using local shard frequencies, not the global frequencies of the word.
  328. The `min_doc_count` criterion is only applied after merging local terms statistics of all shards. In a way the decision to add the
  329. term as a candidate is made without being very _certain_ about if the term will actually reach the required `min_doc_count`.
  330. This might cause many (globally) high frequent terms to be missing in the final result if low frequent but high scoring terms populated
  331. the candidate lists. To avoid this, the `shard_size` parameter can be increased to allow more candidate terms on the shards.
  332. However, this increases memory consumption and network traffic.
  333. `shard_min_doc_count` parameter
  334. The parameter `shard_min_doc_count` regulates the _certainty_ a shard has if the term should actually be added to the candidate list or
  335. not with respect to the `min_doc_count`. Terms will only be considered if their local shard frequency within the set is higher than the
  336. `shard_min_doc_count`. If your dictionary contains many low frequent words and you are not interested in these (for example misspellings),
  337. then you can set the `shard_min_doc_count` parameter to filter out candidate terms on a shard level that will with a reasonable certainty
  338. not reach the required `min_doc_count` even after merging the local frequencies. `shard_min_doc_count` is set to `1` per default and has
  339. no effect unless you explicitly set it.
  340. WARNING: Setting `min_doc_count` to `1` is generally not advised as it tends to return terms that
  341. are typos or other bizarre curiosities. Finding more than one instance of a term helps
  342. reinforce that, while still rare, the term was not the result of a one-off accident. The
  343. default value of 3 is used to provide a minimum weight-of-evidence.
  344. Setting `shard_min_doc_count` too high will cause significant candidate terms to be filtered out on a shard level.
  345. This value should be set much lower than `min_doc_count/#shards`.
  346. ===== Custom background context
  347. The default source of statistical information for background term frequencies is the entire index and this
  348. scope can be narrowed through the use of a `background_filter` to focus in on significant terms within a narrower
  349. context:
  350. [source,js]
  351. --------------------------------------------------
  352. GET news/_search
  353. {
  354. "query" : {
  355. "match" : {
  356. "content" : "madrid"
  357. }
  358. },
  359. "aggs" : {
  360. "tags" : {
  361. "significant_text" : {
  362. "field" : "content",
  363. "background_filter": {
  364. "term" : { "content" : "spain"}
  365. }
  366. }
  367. }
  368. }
  369. }
  370. --------------------------------------------------
  371. // CONSOLE
  372. // TEST[setup:news]
  373. The above filter would help focus in on terms that were peculiar to the city of Madrid rather than revealing
  374. terms like "Spanish" that are unusual in the full index's worldwide context but commonplace in the subset of documents containing the
  375. word "Spain".
  376. WARNING: Use of background filters will slow the query as each term's postings must be filtered to determine a frequency
  377. ===== Dealing with source and index mappings
  378. Ordinarily the indexed field name and the original JSON field being retrieved share the same name.
  379. However with more complex field mappings using features like `copy_to` the source
  380. JSON field(s) and the indexed field being aggregated can differ.
  381. In these cases it is possible to list the JSON _source fields from which text
  382. will be analyzed using the `source_fields` parameter:
  383. [source,js]
  384. --------------------------------------------------
  385. GET news/_search
  386. {
  387. "query" : {
  388. "match" : {
  389. "custom_all" : "elasticsearch"
  390. }
  391. },
  392. "aggs" : {
  393. "tags" : {
  394. "significant_text" : {
  395. "field" : "custom_all",
  396. "source_fields": ["content" , "title"]
  397. }
  398. }
  399. }
  400. }
  401. --------------------------------------------------
  402. // CONSOLE
  403. // TEST[setup:news]
  404. ===== Filtering Values
  405. It is possible (although rarely required) to filter the values for which buckets will be created. This can be done using the `include` and
  406. `exclude` parameters which are based on a regular expression string or arrays of exact terms. This functionality mirrors the features
  407. described in the <<search-aggregations-bucket-terms-aggregation,terms aggregation>> documentation.