 
							pinyin
elasticsearch-analysis-pinyin可以用于繁体字吗?谢谢
回复Elasticsearch • weizhuang 发起了问题 • 2 人关注 • 0 个回复 • 2795 次浏览 • 2018-06-17 12:18
使用ik+pinyin分词后,高亮显示的问题
Elasticsearch • medcl 回复了问题 • 3 人关注 • 1 个回复 • 5961 次浏览 • 2018-06-05 14:55
请教汉字同音词pinyin分词器搜索问题
Elasticsearch • hnj1575565068 回复了问题 • 2 人关注 • 1 个回复 • 4398 次浏览 • 2018-04-27 14:29
elasticsearch拼音提示指定字段问题
Elasticsearch • wengqiankun 回复了问题 • 4 人关注 • 2 个回复 • 4916 次浏览 • 2017-06-21 18:16
elasticsearch-analysis-pinyin更新至es2.4.1和5.0.0-rc1
Elasticsearch • medcl 发表了文章 • 3 个评论 • 4996 次浏览 • 2016-10-13 21:49
curl -XPUT http://localhost:9200/medcl/ -d'
{
    "index" : {
        "analysis" : {
            "analyzer" : {
                "pinyin_analyzer" : {
                    "tokenizer" : "my_pinyin"
                    }
            },
            "tokenizer" : {
                "my_pinyin" : {
                    "type" : "pinyin",
                    "keep_separate_first_letter" : false,
                    "keep_full_pinyin" : true,
                    "keep_original" : false,
                    "limit_first_letter_length" : 16,
                    "lowercase" : true
                }
            }
        }
    }
}'
curl http://localhost:9200/medcl/_a ... lyzer
{
  "tokens" : [ {
    "token" : "liu",
    "start_offset" : 0,
    "end_offset" : 1,
    "type" : "word",
    "position" : 0
  }, {
    "token" : "de",
    "start_offset" : 1,
    "end_offset" : 2,
    "type" : "word",
    "position" : 1
  }, {
    "token" : "hua",
    "start_offset" : 2,
    "end_offset" : 3,
    "type" : "word",
    "position" : 2
  }, {
    "token" : "a",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 3
  }, {
    "token" : "b",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 4
  }, {
    "token" : "c",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 5
  }, {
    "token" : "d",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 6
  }, {
    "token" : "liu",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 7
  }, {
    "token" : "de",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 8
  }, {
    "token" : "hua",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 9
  }, {
    "token" : "wo",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 10
  }, {
    "token" : "bu",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 11
  }, {
    "token" : "zhi",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 12
  }, {
    "token" : "dao",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 13
  }, {
    "token" : "shi",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 14
  }, {
    "token" : "shui",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 15
  }, {
    "token" : "ldhabcdliudehuaw",
    "start_offset" : 0,
    "end_offset" : 16,
    "type" : "word",
    "position" : 16
  } ]
}elasticsearch-analysis-pinyin可以用于繁体字吗?谢谢
回复Elasticsearch • weizhuang 发起了问题 • 2 人关注 • 0 个回复 • 2795 次浏览 • 2018-06-17 12:18
请教汉字同音词pinyin分词器搜索问题
回复Elasticsearch • hnj1575565068 回复了问题 • 2 人关注 • 1 个回复 • 4398 次浏览 • 2018-04-27 14:29
elasticsearch拼音提示指定字段问题
回复Elasticsearch • wengqiankun 回复了问题 • 4 人关注 • 2 个回复 • 4916 次浏览 • 2017-06-21 18:16
elasticsearch-analysis-pinyin更新至es2.4.1和5.0.0-rc1
Elasticsearch • medcl 发表了文章 • 3 个评论 • 4996 次浏览 • 2016-10-13 21:49
curl -XPUT http://localhost:9200/medcl/ -d'
{
    "index" : {
        "analysis" : {
            "analyzer" : {
                "pinyin_analyzer" : {
                    "tokenizer" : "my_pinyin"
                    }
            },
            "tokenizer" : {
                "my_pinyin" : {
                    "type" : "pinyin",
                    "keep_separate_first_letter" : false,
                    "keep_full_pinyin" : true,
                    "keep_original" : false,
                    "limit_first_letter_length" : 16,
                    "lowercase" : true
                }
            }
        }
    }
}'
curl http://localhost:9200/medcl/_a ... lyzer
{
  "tokens" : [ {
    "token" : "liu",
    "start_offset" : 0,
    "end_offset" : 1,
    "type" : "word",
    "position" : 0
  }, {
    "token" : "de",
    "start_offset" : 1,
    "end_offset" : 2,
    "type" : "word",
    "position" : 1
  }, {
    "token" : "hua",
    "start_offset" : 2,
    "end_offset" : 3,
    "type" : "word",
    "position" : 2
  }, {
    "token" : "a",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 3
  }, {
    "token" : "b",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 4
  }, {
    "token" : "c",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 5
  }, {
    "token" : "d",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 6
  }, {
    "token" : "liu",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 7
  }, {
    "token" : "de",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 8
  }, {
    "token" : "hua",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 9
  }, {
    "token" : "wo",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 10
  }, {
    "token" : "bu",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 11
  }, {
    "token" : "zhi",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 12
  }, {
    "token" : "dao",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 13
  }, {
    "token" : "shi",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 14
  }, {
    "token" : "shui",
    "start_offset" : 2,
    "end_offset" : 31,
    "type" : "word",
    "position" : 15
  }, {
    "token" : "ldhabcdliudehuaw",
    "start_offset" : 0,
    "end_offset" : 16,
    "type" : "word",
    "position" : 16
  } ]
}



