步骤:
1、新建job文件
2、将job资源文件打包成zip文件
3、创建project
4、上传zip文件
5、执行Execute Flow -> execute
简单命令
command.job
#command.job
type=command
command=echo 'hello'
打包command.job
工作流
step1.job
# step1.job
type=command
command=echo one
step2.job
# step2.job
type=command
dependencies=step1
command=echo two
打包step1.job与step2.job
mapreduce
wc.job
# wc.job
type=command
command=hadoop jar wc.jar xxx.wordcount /input/wordcount /ouput/wordcount
打包 wc.job 与 wc.jar
hive
hive.job
# hive.job
type=command
command=hive -f 'test.sql'
test.sql
use test;
drop table if exists test;
create table test(id int,name string) row format delimited fields terminated by ',';
load data local inpath 'test.txt' into table aztest;
test.txt
1,java
2,scala
3,hadoop
4,spark
打包这3个文件