Host2NodesMap 是一个学习ReentrantReadWriteLock的一个很好的例子。该类org.apache.hadoop.hdfs.server.namenode.Host2NodesMap用来保存Datanode结点的主机 -> DatanodeDescriptor数组的映射的类,其实一个DatanodeDescriptor中已经包含了Datanode所在主机的字符串的信息,该类设置了一个HashMap<String, DatanodeDescriptor[]>,可见,对于一个Datanode结点,也就是对应的一台主机上,可以存在多个Datanode进程,因此存在一个Datanode主机到一组Datanode进程信息描述的映射。但是一般来说,一个主机上只单独设置一个Datanode进程。通过上面叙述,可以想到,作为一个Host2NodesMap类,应该提供向map映射表中添加<Host, DatanodeDescriptor[]>对的操作,删除指定Host对应的DatanodeDescriptor[],获取到指定Host对应的DatanodeDescriptor[]信息,等等。
package org.apache.hadoop.hdfs.server.namenode;
import java.util.*;
import java.util.concurrent.locks.ReadWriteLock;
import java.util.concurrent.locks.ReentrantReadWriteLock;
class Host2NodesMap {
private HashMap<String, DatanodeDescriptor[]> map
= new HashMap<String, DatanodeDescriptor[]>();
private Random r = new Random();
private ReadWriteLock hostmapLock = new ReentrantReadWriteLock();
/** Check if node is already in the map. */
boolean contains(DatanodeDescriptor node) {
if (node==null) {
return false;
}
String host = node.getHost();
hostmapLock.readLock().lock();
try {
DatanodeDescriptor[] nodes = map.get(host);
if (nodes != null) {
for(DatanodeDescriptor containedNode:nodes) {
if (node==containedNode) {
return true;
}
}
}
} finally {
hostmapLock.readLock().unlock();
}
return false;
}
/** add node to the map
* return true if the node is added; false otherwise.
*/
boolean add(DatanodeDescriptor node) {
hostmapLock.writeLock().lock();
try {
if (node==null || contains(node)) {
return false;
}
String host = node.getHost();
DatanodeDescriptor[] nodes = map.get(host);
DatanodeDescriptor[] newNodes;
if (nodes==null) {
newNodes = new DatanodeDescriptor[1];
newNodes[0]=node;
} else { // rare case: more than one datanode on the host
newNodes = new DatanodeDescriptor[nodes.length+1];
System.arraycopy(nodes, 0, newNodes, 0, nodes.length);
newNodes[nodes.length] = node;
}
map.put(host, newNodes);
return true;
} finally {
hostmapLock.writeLock().unlock();
}
}
/** remove node from the map
* return true if the node is removed; false otherwise.
*/
boolean remove(DatanodeDescriptor node) {
if (node==null) {
return false;
}
String host = node.getHost();
hostmapLock.writeLock().lock();
try {
DatanodeDescriptor[] nodes = map.get(host);
if (nodes==null) {
return false;
}
if (nodes.length==1) {
if (nodes[0]==node) {
map.remove(host);
return true;
} else {
return false;
}
}
//rare case
int i=0;
for(; i<nodes.length; i++) {
if (nodes[i]==node) {
break;
}
}
if (i==nodes.length) {
return false;
} else {
DatanodeDescriptor[] newNodes;
newNodes = new DatanodeDescriptor[nodes.length-1];
System.arraycopy(nodes, 0, newNodes, 0, i);
System.arraycopy(nodes, i+1, newNodes, i, nodes.length-i-1);
map.put(host, newNodes);
return true;
}
} finally {
hostmapLock.writeLock().unlock();
}
}
/** get a data node by its host.
* @return DatanodeDescriptor if found; otherwise null.
*/
DatanodeDescriptor getDatanodeByHost(String host) {
if (host==null) {
return null;
}
hostmapLock.readLock().lock();
try {
DatanodeDescriptor[] nodes = map.get(host);
// no entry
if (nodes== null) {
return null;
}
// one node
if (nodes.length == 1) {
return nodes[0];
}
// more than one node
return nodes[r.nextInt(nodes.length)];
} finally {
hostmapLock.readLock().unlock();
}
}
/**
* Find data node by its name.
*
* @return DatanodeDescriptor if found or null otherwise
*/
public DatanodeDescriptor getDatanodeByName(String name) {
if (name==null) {
return null;
}
int colon = name.indexOf(":");
String host;
if (colon < 0) {
host = name;
} else {
host = name.substring(0, colon);
}
hostmapLock.readLock().lock();
try {
DatanodeDescriptor[] nodes = map.get(host);
// no entry
if (nodes== null) {
return null;
}
for(DatanodeDescriptor containedNode:nodes) {
if (name.equals(containedNode.getName())) {
return containedNode;
}
}
return null;
} finally {
hostmapLock.readLock().unlock();
}
}
}