大数据相关参考

本文汇总了大数据处理的相关资源,包括Spark Streaming与Redis的整合、Maven与Ant的配合使用、ECharts图表库的应用、Spark读取配置文件的方法、Hive基于日期的分区策略、CDH高可用设置、Azkaban的安装与实战、Oozie的使用体验以及Kafka、HDFS、Hive之间的数据流动。同时推荐了Spark优化参考书籍,并提供了IDEA搭建Spark开发环境的教程。
摘要由CSDN通过智能技术生成

sparkstreaming写redis : https://blog.csdn.net/qq_28666339/article/details/79234301

maven-ant-run copy tasks: http://ant.apache.org/manual/Tasks/copy.html

echarts: https://gallery.echartsjs.com/explore.html#sort=rank~timeframe=all~author=all

https://blog.csdn.net/qq_36275889/article/details/83383150

echarts3实现暂无数据:https://github.com/apache/incubator-echarts/issues/4829

spark读取配置文件:

nohup spark2-submit --master yarn --num-executors 2 --executor-cores 1 --executor-memory 3G --driver-memory 1G --class com.marfosec.ods.OdsAccessOutDetail --files config.properties bigdata-cernetbam-1.0.0.jar 1 &
val props : Properties = new Properties
if (isCluster == 1) {
    props.load(new FileInputStream("config.properties"))
} else {
    val resource: URL = this.getClass.getClassLoader.getResource("config.properties")
    val in:InputStream = new FileInputStream(resource.getPath)
    props.load(in)
}

hive基于日期分区:https://blog.csdn.net/dylanzr/article/details/86187552

cdh高可用:https://blog.csdn.net/u011142688/article/details/82078132

azkaban安装:https://blog.csdn.net/weixin_42179685/article/details/90716366

https://blog.csdn.net/hg_harvey/article/details/80342396

问题解决:https://yq.aliyun.com/articles/648399

azkaban实战:https://blog.csdn.net/tototuzuoquan/article/details/73251616

https://www.jianshu.com/p/01188607a794?nomobile=yes

cdh oozie:

https://blog.csdn.net/hxiaowang/article/details/78551106 

https://blog.csdn.net/qq_24908345/article/details/80017660

(oozie真是太难用了, 推荐使用azkaban)

spark优化参考书:https://www.jb51.net/books/612370.html

spark消费kafka数据,入到hive中:https://blog.csdn.net/u012164361/article/details/79742201

java, kafka写数据到hdfs:https://blog.csdn.net/u013385018/article/details/80689546

cdh版本的maven依赖:https://blog.csdn.net/hexinghua0126/article/details/80292905

cdh安装spark2.3:https://blog.csdn.net/lichangzai/article/details/82225494

https://www.jianshu.com/p/170ffe85c063/

https://www.cnblogs.com/zengxiaoliang/p/6478859.html(spark2.1.0)

https://blog.csdn.net/greenplum_xiaofan/article/details/97622256

https://yq.aliyun.com/articles/341408

https://blog.csdn.net/liu16659/article/details/81074465

idea搭建spark开发环境:https://blog.csdn.net/yiluohan0307/article/details/79568363

scala和java实现wordcount:https://www.cnblogs.com/byrhuangqiang/p/4017725.html(java在spark上非常繁琐,建议学scala)

kafka查看消息解压情况:https://blog.51cto.com/12473494/2420105

spark2-shell读取hive数据:https://www.cnblogs.com/xinfang520/p/7985939.html

sqoop将mysql导入到hive:https://www.cnblogs.com/xuyou551/p/7998846.html

 

 

 

 

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值