下载地址,选择对应版本https://github.com/medcl/elasticsearch-analysis-ik/releases
安装ik分词器
在 elasticsearch 目录中的 \plugins 目录下新建 ik 目录,然后将下载完成的 elasticsearch-analysis-ik 解压至此目录下。或者将压缩后的文件复制到ik目录下,分词器到这里就安装成功了
如果elasticsearch版本是6.x以下的。貌似要进入elasticsearch目录,进入config配置文件目录修改elasticsearch.yml,添加如下配置。6.x及以上不需要配置,配置反而启动不了,会报错
index.analysis.analyzer.ik.type : "ik"
重启elasticsearch
如果是集群有其他节点的话,其他节点同样的操作。安装步骤同样适用于linux,只是插件需要是linux的,或者在线打包安装
测试基本分词:
"analyzer":"ik_smart"
ik_smart:最粗粒度的分词
ik_max_word:最细粒度的分词
如果是"analyzer":"ik_max_word"最细粒度的就很多了:
{
- "tokens": [
- {
- "token": "中华人民共和国",
- "start_offset": 0,
- "end_offset": 7,
- "type": "CN_WORD",
- "position": 0
- ,
- {
- "token": "中华人民",
- "start_offset": 0,
- "end_offset": 4,
- "type": "CN_WORD",
- "position": 1
- ,
- {
- "token": "中华",
- "start_offset": 0,
- "end_offset": 2,
- "type": "CN_WORD",
- "position": 2
- ,
- {
- "token": "华人",
- "start_offset": 1,
- "end_offset": 3,
- "type": "CN_WORD",
- "position": 3
- ,
- {
- "token": "人民共和国",
- "start_offset": 2,
- "end_offset": 7,
- "type": "CN_WORD",
- "position": 4
- ,
- {
- "token": "人民",
- "start_offset": 2,
- "end_offset": 4,
- "type": "CN_WORD",
- "position": 5
- ,
- {
- "token": "共和国",
- "start_offset": 4,
- "end_offset": 7,
- "type": "CN_WORD",
- "position": 6
- ,
- {
- "token": "共和",
- "start_offset": 4,
- "end_offset": 6,
- "type": "CN_WORD",
- "position": 7
- ,
- {
- "token": "国",
- "start_offset": 6,
- "end_offset": 7,
- "type": "CN_CHAR",
- "position": 8
- ,
- {
- "token": "国歌",
- "start_offset": 7,
- "end_offset": 9,
- "type": "CN_WORD",
- "position": 9
- {
}
查询book带有相关关键词的数据