Database Reference
In-Depth Information
3. Create Hadoop temporary directory
On the Linux file system, you create a Hadoop temporary directory, as shown below. This will give Hadoop a working
area. Set the ownership to the hadoop user and also set the directory permissions:
[root@hc1nn local]# mkdir -p /app/hadoop/tmp
[root@hc1nn local]# chown -R hadoop:hadoop /app/hadoop
[root@hc1nn local]# chmod 750 /app/hadoop/tmp
4. Set up conf/core-site.xml
You set up the configuration for the Hadoop core component. This file configuration is based on XML; it defines the
Hadoop temporary directory and default file system access. There are many more options that can be specified; see
the Hadoop site (hadoop.apache.org ) for details.
Add the following text to the file between the configuration tags:
<property>
<name>hadoop.tmp.dir</name>
<value>/app/hadoop/tmp</value>
<description>A base for other temporary directories.</description>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://localhost:54310</value>
<description>The name of the default file system.</description>
</property>
5. Set up conf/mapred-site.xml
Next, you set up the basic configuration for the Map Reduce component, adding the following between the
configuration tags. This defines the host and port name for each Job Tracker server.
<property>
<name>mapred.job.tracker</name>
<value>localhost:54311</value>
<description>The host and port for the Map Reduce job tracker
</description>
</property>
<property>
<name>mapred.job.tracker.http.address</name>
<value>localhost:50030</value>
</property>
<property>
<name>mapred.task.tracker.http.address</name>
<value>localhost:50060</value>
</property>
 
Search WWH ::




Custom Search