Elasticsearch集群数据批量导入

本文转自http://www.jianshu.com/u/e02df63eaa87

1、数据形式

我们采用Person的作为数据出发点,将文件中的序列化为Json的Person对象导入Elasticsearch集群中。
本文中的代码详见:https://github.com/hawkingfoo/es-batch-import

1.1 数据类型

public class Person {
    private int pid;            // person id
    private int age;
    private boolean sex;
    private String name;
    private String addr;
}
1.2 序列化Json后的文件类型

Person.dat id与json串以\t作为分割。

 
0 {"pid":0,"age":41,"sex":true,"name":"Lucy","addr":"Shanghai"}
1 {"pid":1,"age":9,"sex":true,"name":"Jenny","addr":"Shenzhen"} 
2 {"pid":2,"age":9,"sex":true,"name":"Lily","addr":"Tianjin"} 
3 {"pid":3,"age":42,"sex":false,"name":"David","addr":"Guangzhou"}
4 {"pid":4,"age":40,"sex":true,"name":"Mary","addr":"Chongqing"} 
5 {"pid":5,"age":3,"sex":true,"name":"Jenny","addr":"Guangzhou"}
2、ES建立index和mapping

建立5个分片1个副本的index,其中ES的type为infos,对应的mapping如下:

{ "settings": { "index": { "creation_date": "1470300617555", "legacy": { "routing": { "hash": { "type": "org.elasticsearch.cluster.routing.DjbHashFunction" }, "use_type": "false" } }, "number_of_shards": "5", "number_of_replicas": "1", "uuid": "mJXGBmnYS12mXBo0aGrR3Q", "version": { "created": "1070099", "upgraded": "2030499" } } }, "mappings": { "infos": { "_timestamp": {}, "properties": { "sex": { "type": "boolean" }, "name": { "index": "not_analyzed", "type": "string" }, "pid": { "type": "integer" }, "addr": { "index": "not_analyzed", "type": "string" }, "age": { "type": "integer" } } } } }


3、导入程序模块

3.1 流程图





整个导入模块的流程图如上,Main创建ESClientBulkProcessor;读取Person.dat中的Json串,组成UpdateRequest后加入到BulkProcessor中,当BulkProcessor满足一定的写入条件后,会批量进行发送到ES集群。

3.2 ESClient建立

添加Maven依赖:

<dependency>
        <groupId>org.elasticsearch</groupId>
        <artifactId>elasticsearch</artifactId>
        <version>2.3.4</version>
</dependency>
// ESConfig 
public class ESConfig { private String esClusterName; // 集群名称 
private String esClusterAddress; // 集群地址
private String esIndex; // ES库 
private String esType; // ES表 
private int batchSize; // 批量导入大小 
private String filePath; // 导入文件的路径 
private int esThreadNum; // 导入到ES的并发数量 
private String localClientIP; // 本机IP地址 
public String getEsClusterName() { return esClusterName; } public ESConfig setEsClusterName(String esClusterName) { this.esClusterName = esClusterName; return this; } public String getEsClusterAddress() { return esClusterAddress; } public ESConfig setEsClusterAddress(String esClusterAddress) { this.esClusterAddress = esClusterAddress; return this; } public String getEsIndex() { return esIndex; } public ESConfig setEsIndex(String esIndex) { this.esIndex = esIndex; return this; } public String getEsType() { return esType; } public ESConfig setEsType(String esType) { this.esType = esType; return this; } public int getBatchSize() { return batchSize; } public ESConfig setBatchSize(int batchSize) { this.batchSize = batchSize; return this; } public String getFilePath() { return filePath; } public ESConfig setFilePath(String filePath) { this.filePath = filePath; return this; } public int getEsThreadNum() { return esThreadNum; } public ESConfig setEsThreadNum(int esThreadNum) { this.esThreadNum = esThreadNum; return this; } public String getLocalClientIP() { return localClientIP; } public ESConfig setLocalClientIP(String localClientIP) { this.localClientIP = localClientIP; return this; } }

在3.1节中,我们曾提到过满足发送条件这个概念,对应于上面BulkProcessor中的3个set方法。分别是:

  • 当导入数据(UpdateRequest)的个数达到后,进行发送;
  • 当导入数据的大小达到1MB后,进行发送;
  • 当距离上一次发送超过60秒时,进行发送。
3.3 读取并组装UpdateRequest

ESImporter:

public class ESImporter { private static final Logger logger = LogManager.getLogger(ESImporter.class); public void importer(ESConfig esConfig) { File file = new File(esConfig.getFilePath()); BufferedReader reader = null; // 创建BulkProcessor BulkProcessor bulkProcessor = new ESClient().createBulkProcessor(esConfig); if (bulkProcessor == null) { logger.error("create bulk processor failed."); return; } UpdateRequest updateRequest; String[] arrStr; try { reader = new BufferedReader(new FileReader(file)); String tempString; // 一次读入一行,直到读入null为文件结束 while ((tempString = reader.readLine()) != null) { arrStr = tempString.split("\t"); if (arrStr.length != 2) { continue; } updateRequest = new UpdateRequest(esConfig.getEsIndex(), esConfig.getEsType(), arrStr[0]) .doc(arrStr[1]).docAsUpsert(true); bulkProcessor.add(updateRequest); } reader.close(); } catch (Exception e) { e.printStackTrace(); } finally { try { if (reader != null) { reader.close(); } if (bulkProcessor != null) { bulkProcessor.awaitClose(1, TimeUnit.MINUTES); } } catch (Exception e) { // do nothing } } } }

这个模块主要是读取文件中的Json行,组装成UpdateRequest后,加入到bulkProcessor中。

3.4 服务启动模块

ImportMain:

public class ImportMain { private static final Logger logger = LogManager.getLogger(ImportMain.class); public static void main(String[] args) { try { if (args.length < 1) { System.err.println("usage: <file_path>"); System.exit(1); } ESConfig esConfig = new ESConfig() .setEsClusterName("elasticsearch") .setEsClusterAddress("127.0.0.1:9300") .setEsIndex("person") .setEsType("infos") .setBatchSize(100) .setFilePath(args[0]) .setEsThreadNum(1); long begin = System.currentTimeMillis(); ESImporter esImporter = new ESImporter(); esImporter.importer(esConfig); long cost = System.currentTimeMillis() - begin; logger.info("import end. cost:[{}ms]", cost); } catch (Exception e) { logger.error("exception:", e); } } }
3.5 代码目录

3.6 ES集群查看
导入结束后,在ES集群上可以看到导入的docs。








  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值