N(逆天邪神)

  本篇文章为你整理了N(逆天邪神)的详细内容,包含有nba在线观看免费观看 逆天邪神 nba赛程 nba直播吧 N,希望能帮助你了解 N。

  IMPORTANT: No additional bug fixes or documentation updates

  will be released for this version. For the latest information, see the

  current release documentation.

  
N-gram tokenizeredit

  The ngram tokenizer first breaks text down into words whenever it encounters

  one of a list of specified characters, then it emits

  N-grams of each word of the specified

  length.

  N-grams are like a sliding window that moves across the word - a continuous

  sequence of characters of the specified length. They are useful for querying

  languages that don t use spaces or that have long compound words, like German.

  Example outputedit

  With the default settings, the ngram tokenizer treats the initial text as a

  single token and produces N-grams with minimum length 1 and maximum length

  2:

  

POST _analyze

 

   "tokenizer": "ngram",

   "text": "Quick Fox"

  }

 

  
Character classes that should be included in a token. Elasticsearch

  will split on characters that don t belong to the classes specified.

  Defaults to [] (keep all characters).

  Character classes may be any of the following:

  
Custom characters that should be treated as part of a token. For example,

  setting this to +-_ will make the tokenizer treat the plus, minus and

  underscore sign as part of a token.

  
It usually makes sense to set min_gram and max_gram to the same

  value. The smaller the length, the more documents will match but the lower

  the quality of the matches. The longer the length, the more specific the

  matches. A tri-gram (length 3) is a good place to start.

  
The index level setting index.max_ngram_diff controls the maximum allowed

  difference between max_gram and min_gram.

  Example configurationedit

  In this example, we configure the ngram tokenizer to treat letters and

  digits as tokens, and to produce tri-grams (grams of length 3):

  

PUT my_index

 

   "settings": {

   "analysis": {

   "analyzer": {

   "my_analyzer": {

   "tokenizer": "my_tokenizer"

   "tokenizer": {

   "my_tokenizer": {

   "type": "ngram",

   "min_gram": 3,

   "max_gram": 3,

   "token_chars": [

   "letter",

   "digit"

  POST my_index/_analyze

   "analyzer": "my_analyzer",

   "text": "2 Quick Foxes."

  }

 

  以上就是N(逆天邪神)的详细内容,想要了解更多 N的内容,请持续关注盛行IT软件开发工作室。

郑重声明:本文由网友发布,不代表盛行IT的观点,版权归原作者所有,仅为传播更多信息之目的,如有侵权请联系,我们将第一时间修改或删除,多谢。

留言与评论(共有 条评论)
   
验证码: