ccah-500 第7题 swap Hadoop daemon data from RAM to disk

原创 2016年06月01日 17:36:30

7. You want to node to only swap Hadoop daemon data from RAM to disk when absolutely necessary. What should you do? 

A. Delete the /dev/vmswap file on the node 

B. Delete the /etc/swap file on the node 

C. Set the ram.swap parameter to 0 in core-site.xml 

D. Set vm.swappiness=0 in /etc/sysctl.conf

E. Delete the /swapfile file on the node 

Answer: D 



Improving Performance
This section summarizes some recent code improvements and configuration best practices.

Setting the vm.swappiness Linux Kernel Parameter
vm.swappiness is a Linux Kernel Parameter that controls how aggressively memory pages are swapped to disk. It can be set to a value between 0-100; the higher the value, the more aggressive the kernel is in seeking out inactive memory pages and swapping them to disk.

You can see what value vm.swappiness is currently set to by looking at /proc/sys/vm; for example:

cat /proc/sys/vm/swappiness
On most systems, it is set to 60 by default. This is not suitable for Hadoop clusters nodes, because it can cause processes to get swapped out even when there is free memory available. This can affect stability and performance, and may cause problems such as lengthy garbage collection pauses for important system daemons. Cloudera recommends that you set this parameter to 0; for example:

# sysctl -w vm.swappiness=0


ccah-500 第9题 How would you tune your io.sort.mb value to achieve maximum memory to disk I/O ratio?

9.You observed that the number of spilled records from Map tasks far exceeds the number of map out...

ccah-500 第45题 You want to minimize the chance of data loss in your cluster. What should you do

45.You have A 20 node Hadoop cluster, with 18 slave nodes and 2 master nodes running HDFS High Avail...

ccah-500 第27题 where does the Mapper place the intermediate data of each Map Task

27.During the execution of a MapReduce v2 (MRv2) job on YARN, where does the Mapper place the interm...

ccah-500 第13题Which three basic configuration parameters must you set to migrate

13.Which three basic configuration parameters must you set to migrate your cluster from MapReduce 1 ...

ccah-500 第47题 What is the purpose of ZooKeeper in such a configuration

47.You decide to create a cluster which runs HDFS in High Availability mode with automatic failover ...

ccah-500 第40题 maintain your MRv1 TaskTracker slot capacities when you migrate. What should you do

40.You are migrating a cluster from MApReduce version 1 (MRv1) to MapReduce version 2(MRv2) on YARN....

ccah-500 第49题 What occurs when you execute the command: hdfs haadmin -failover nn01 nn02

49.Your cluster implements HDFS High Availability (HA). Your two NameNodes are named nn01 and nn02. ...

ccah-500 第32题 a new user on the cluster can submit jobs into their own queue application submission

32.Your cluster is running MapReduce version 2 (MRv2) on YARN. Your ResourceManager is configured to...

ccah-500 第57题 Which two best describes how FIFO Scheduler arbitrates the cluster resources for job

57.You have a cluster running with a FIFO scheduler enabled. You submit a large job A to the cluster...

How To Recover From Corrupted OCR Disk

It is very common where a DBA is left with corrupted OCR disk without having any good backup. The s...
您举报文章:ccah-500 第7题 swap Hadoop daemon data from RAM to disk