大数据之YARN

yarn: 运算资源(jar包、内存、cpu、配置文件)调度系统。

一、mapreduce提交job运行在yarn流程:
在这里插入图片描述
yarn的调度机制有:队列(先进先出)、fair、capacity;

yarn只负责程序运行所需资源的分配回收等调度任务,与应用程序的内部运行机制无关,所以yarn成为了一个通用的资源调度平台。比如mr/spark/storm,都可以借助它来实现。

注: hadoop1没有yarn的概念,但是有一个jobTracker和TaskTracher两种角色。

二、使用java -jar方式运行mr程序(正常情况是用hadoop jar方式)

1、下载四个配置文件到项目里
core-site.xml、 hdfs-site.xml 、mapred-site.xml 、yarn-site.xml

2、指定jar所在的本地位置

job.setJar("/home/hadoop/wc.jar");

三、本地运行yarn集群

1、指定jar所在的本地位置

job.setJar("c:/wc.jar");

2、配置运行参数
在这里插入图片描述
3、下载YARNRunner.java到本地

4、本地启动

四、reduce端join算法实现

以mapreduce程序实现订单表和商品表查询。

1、创建订单-商品实体类

	public class InfoBean implements Writable {
	
		private int order_id;
		private String dateString;
		private String p_id;
		private int amount;
		private String pname;
		private int category_id;
		private float price;
	
		// flag=0表示这个对象是封装订单表记录
		// flag=1表示这个对象是封装产品信息记录
		private String flag;
	
		public InfoBean() {
		}
	
		public void set(int order_id, String dateString, String p_id, int amount, String pname, int category_id, float price, String flag) {
			this.order_id = order_id;
			this.dateString = dateString;
			this.p_id = p_id;
			this.amount = amount;
			this.pname = pname;
			this.category_id = category_id;
			this.price = price;
			this.flag = flag;
		}
	
		public int getOrder_id() {
			return order_id;
		}
	
		public void setOrder_id(int order_id) {
			this.order_id = order_id;
		}
	
		public String getDateString() {
			return dateString;
		}
	
		public void setDateString(String dateString) {
			this.dateString = dateString;
		}
	
		public String getP_id() {
			return p_id;
		}
	
		public void setP_id(String p_id) {
			this.p_id = p_id;
		}
	
		public int getAmount() {
			return amount;
		}
	
		public void setAmount(int amount) {
			this.amount = amount;
		}
	
		public String getPname() {
			return pname;
		}
	
		public void setPname(String pname) {
			this.pname = pname;
		}
	
		public int getCategory_id() {
			return category_id;
		}
	
		public void setCategory_id(int category_id) {
			this.category_id = category_id;
		}
	
		public float getPrice() {
			return price;
		}
	
		public void setPrice(float price) {
			this.price = price;
		}
	
		public String getFlag() {
			return flag;
		}
	
		public void setFlag(String flag) {
			this.flag = flag;
		}
	
		/**
		 * private int order_id; private String dateString; private int p_id;
		 * private int amount; private String pname; private int category_id;
		 * private float price;
		 */
		@Override
		public void write(DataOutput out) throws IOException {
			out.writeInt(order_id);
			out.writeUTF(dateString);
			out.writeUTF(p_id);
			out.writeInt(amount);
			out.writeUTF(pname);
			out.writeInt(category_id);
			out.writeFloat(price);
			out.writeUTF(flag);
	
		}
	
		@Override
		public void readFields(DataInput in) throws IOException {
			this.order_id = in.readInt();
			this.dateString = in.readUTF();
			this.p_id = in.readUTF();
			this.amount = in.readInt();
			this.pname = in.readUTF();
			this.category_id = in.readInt();
			this.price = in.readFloat();
			this.flag = in.readUTF();
	
		}

	@Override
	public String toString() {
		return "order_id=" + order_id + ", dateString=" + dateString + ", p_id=" + p_id + ", amount=" + amount + ", pname=" + pname + ", category_id=" + category_id + ", price=" + price + ", flag=" + flag;
	}
}

2、创建运行类

package cn.itcast.bigdata.mr.rjoin;

import java.io.IOException;
import java.util.ArrayList;

import org.apache.commons.beanutils.BeanUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.input.FileSplit;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;

/**
 * 订单表和商品表合到一起
order.txt(订单id, 日期, 商品编号, 数量)
	1001	20150710	P0001	2
	1002	20150710	P0001	3
	1002	20150710	P0002	3
	1003	20150710	P0003	3
product.txt(商品编号, 商品名字, 价格, 数量)
	P0001	小米5	1001	2
	P0002	锤子T1	1000	3
	P0003	锤子	1002	4
 */
public class RJoin {

	static class RJoinMapper extends Mapper<LongWritable, Text, Text, InfoBean> {
		InfoBean bean = new InfoBean();
		Text k = new Text();

		@Override
		protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {

			String line = value.toString();

			FileSplit inputSplit = (FileSplit) context.getInputSplit();
			String name = inputSplit.getPath().getName();
			// 通过文件名判断是哪种数据
			String pid = "";
			if (name.startsWith("order")) {
				String[] fields = line.split("\t");
				// id date pid amount
				pid = fields[2];
				bean.set(Integer.parseInt(fields[0]), fields[1], pid, Integer.parseInt(fields[3]), "", 0, 0, "0");

			} else {
				String[] fields = line.split("\t");
				// id pname category_id price
				pid = fields[0];
				bean.set(0, "", pid, 0, fields[1], Integer.parseInt(fields[2]), Float.parseFloat(fields[3]), "1");

			}
			k.set(pid);
			context.write(k, bean);
		}

	}

	static class RJoinReducer extends Reducer<Text, InfoBean, InfoBean, NullWritable> {

		@Override
		protected void reduce(Text pid, Iterable<InfoBean> beans, Context context) throws IOException, InterruptedException {
			InfoBean pdBean = new InfoBean();
			ArrayList<InfoBean> orderBeans = new ArrayList<InfoBean>();

			for (InfoBean bean : beans) {
				if ("1".equals(bean.getFlag())) {	//产品的
					try {
						BeanUtils.copyProperties(pdBean, bean);
					} catch (Exception e) {
						e.printStackTrace();
					}
				} else {
					InfoBean odbean = new InfoBean();
					try {
						BeanUtils.copyProperties(odbean, bean);
						orderBeans.add(odbean);
					} catch (Exception e) {
						e.printStackTrace();
					}
				}

			}

			// 拼接两类数据形成最终结果
			for (InfoBean bean : orderBeans) {

				bean.setPname(pdBean.getPname());
				bean.setCategory_id(pdBean.getCategory_id());
				bean.setPrice(pdBean.getPrice());

				context.write(bean, NullWritable.get());
			}
		}
	}

	public static void main(String[] args) throws Exception {
		Configuration conf = new Configuration();
		Job job = Job.getInstance(conf);

		// 指定本程序的jar包所在的本地路径
		// job.setJarByClass(RJoin.class);
        job.setJar("c:/join.jar");

		//job.setJarByClass(RJoin.class);
		// 指定本业务job要使用的mapper/Reducer业务类
		job.setMapperClass(RJoinMapper.class);
		job.setReducerClass(RJoinReducer.class);

		// 指定mapper输出数据的kv类型
		job.setMapOutputKeyClass(Text.class);
		job.setMapOutputValueClass(InfoBean.class);

		// 指定最终输出的数据的kv类型
		job.setOutputKeyClass(InfoBean.class);
		job.setOutputValueClass(NullWritable.class);

		// 指定job的输入原始文件所在目录
		FileInputFormat.setInputPaths(job, new Path(args[0]));
		// 指定job的输出结果所在目录
		FileOutputFormat.setOutputPath(job, new Path(args[1]));

		// 将job中配置的相关参数,以及job所用的java类所在的jar包,提交给yarn去运行
		/* job.submit(); */
		boolean res = job.waitForCompletion(true);
		System.exit(res ? 0 : 1);

	}
}

缺点:这种方式中,join的操作是在reduce阶段完成,reduce端的处理压力太大,map节点的运算负载则很低,资源利用率不高,且在reduce阶段极易产生数据倾斜。

五、map端实现join算法(本地运行)

把产品信息文件通过DistributedCache,发给每台maptask机器上。

在这里插入图片描述
编写MapSideJoin

	package cn.itcast.bigdata.mr.mapsidejoin;
	
	import java.io.BufferedReader;
	import java.io.FileInputStream;
	import java.io.IOException;
	import java.io.InputStreamReader;
	import java.net.URI;
	import java.util.HashMap;
	import java.util.Map;
	import org.apache.commons.lang.StringUtils;
	import org.apache.hadoop.conf.Configuration;
	import org.apache.hadoop.fs.Path;
	import org.apache.hadoop.io.LongWritable;
	import org.apache.hadoop.io.NullWritable;
	import org.apache.hadoop.io.Text;
	import org.apache.hadoop.mapreduce.Job;
	import org.apache.hadoop.mapreduce.Mapper;
	import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
	import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
	
	public class MapSideJoin {
	
		public static class MapSideJoinMapper extends Mapper<LongWritable, Text, Text, NullWritable> {
			// 用一个hashmap来加载保存产品信息表
			Map<String, String> pdInfoMap = new HashMap<String, String>();
	
			Text k = new Text();
	
			/**
			 * 通过阅读父类Mapper的源码,发现 setup方法是在maptask处理数据之前调用一次 可以用来做一些初始化工作
			 */
			@Override
			protected void setup(Context context) throws IOException, InterruptedException {
				BufferedReader br = new BufferedReader(new InputStreamReader(new FileInputStream("pdts.txt")));
				String line;
				while (StringUtils.isNotEmpty(line = br.readLine())) {
					String[] fields = line.split(",");
					pdInfoMap.put(fields[0], fields[1]);
				}
				br.close();
			}
	
			// 由于已经持有完整的产品信息表,所以在map方法中就能实现join逻辑了
			@Override
			protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
				String orderLine = value.toString();
				String[] fields = orderLine.split("\t");
				String pdName = pdInfoMap.get(fields[1]);
				k.set(orderLine + "\t" + pdName);
				context.write(k, NullWritable.get());
			}
	
		}
	
		public static void main(String[] args) throws Exception {
	
			Configuration conf = new Configuration();
	
			Job job = Job.getInstance(conf);
	
			job.setJarByClass(MapSideJoin.class);
	
			job.setMapperClass(MapSideJoinMapper.class);
			job.setOutputKeyClass(Text.class);
			job.setOutputValueClass(NullWritable.class);
	
			FileInputFormat.setInputPaths(job, new Path("D:/srcdata/mapjoininput"));
			FileOutputFormat.setOutputPath(job, new Path("D:/temp/output"));
	
			// 指定需要缓存一个文件到所有的maptask运行节点工作目录
			/* job.addArchiveToClassPath(archive); */// 缓存jar包到task运行节点的classpath中
			/* job.addFileToClassPath(file); */// 缓存普通文件到task运行节点的classpath中
			/* job.addCacheArchive(uri); */// 缓存压缩包文件到task运行节点的工作目录
			/* job.addCacheFile(uri) */// 缓存普通文件到task运行节点的工作目录
	
			// 将产品表文件缓存到task工作节点的工作目录中去
			job.addCacheFile(new URI("file:/D:/srcdata/mapjoincache/pdts.txt"));
	
			//map端join的逻辑不需要reduce阶段,设置reducetask数量为0
			job.setNumReduceTasks(0);
			
			boolean res = job.waitForCompletion(true);
			System.exit(res ? 0 : 1);
	
		}
	
	}
  • 1
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

大数据同盟会

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值