Database Reference
In-Depth Information
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
<value>hc2nn:8031</value>
</property>
<property>
<name>yarn.resourcemanager.admin.address</name>
<value>hc2nn:8033</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address</name>
<value>hc2nn:8088</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.class</name>
<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler</value>
</property>
I place the following entries into the slaves file to indicate where the slave data nodes reside:
hc2r1m1
hc2r1m2
hc2r1m3
hc2r1m4
To ensure that the Hadoop scheduler is correctly configured, I copy the fair-scheduler.xml file that I created in
Chapter 5 and place it in the configuration directory /etc/hadoop/conf. Now I create the Linux and HDFS file system
locations that are needed by the Hadoop server, and I set their ownership and permissions:
[root@hc1nn conf]# mkdir -p /var/lib/hadoop-hdfs/cache/hdfs/dfs/name
[root@hc1nn conf]# chown -R hdfs:hdfs /var/lib/hadoop-hdfs/cache/hdfs/dfs/name
[root@hc1nn conf]# chmod 700 /var/lib/hadoop-hdfs/cache/hdfs/dfs/name
I set the JAVA_HOME variable so that the Hadoop server knows where to find Java:
export JAVA_HOME=/usr/lib/jvm/java-1.6.0-openjdk.x86_64
As the hdfs user, I format the Hadoop file system without the Hadoop servers running:
[root@hc2nn conf]# su - hdfs
-bash-4.1$ hdfs namenode -format
Next, I create the directories used by YARN, setting the ownership and group membership to the YARN Linux
user and group:
[root@hc1nn conf]# mkdir -p /var/log/hadoop-yarn/containers
[root@hc1nn conf]# mkdir -p /var/log/hadoop-yarn/apps
[root@hc1nn conf]# chown -R yarn:yarn /var/log/hadoop-yarn/containers
[root@hc1nn conf]# chown -R yarn:yarn /var/log/hadoop-yarn/apps
[root@hc1nn conf]# chmod 755 /var/log/hadoop-yarn/containers
[root@hc1nn conf]# chmod 755 /var/log/hadoop-yarn/apps
 
Search WWH ::




Custom Search