Skip to content
Advertisement

Linux and Hadoop : Mounting a disk and increasing the cluster capacity [closed]

First of all, I’m a total noob at hadoop and linux.I have a cluster of five nodes , which when starts shows a each node capacity only 46.6 GB while each machine had around 500 gb space which i dont know how to allocate to these nodes.

(1) Do I have to change the datanode and namenode file size(i checked these and it shows the same space remaining as in the Datanode information tab)? if so how should i do that.

(2)Also this 500gb disk is only shown when i do $lsblk command and not when i do $df -H command. Does that mean its not mounted? These are the results of the commands. Can someone explain what does this mean..

JavaScript

Please help. Thanks in advance. First can someone help me understand why its showing different memory disks and what it means and where does it reside ?! I seem to not able to figure it out

Advertisement

Answer

You are right. Your second disk (sdb) is not mounted anywhere. If you are going to dedicate the whole disk to hadoop data, here is how you should format and mount it:

Format your disk:

JavaScript

For mounting edit the file /etc/fstab. Add this line:

JavaScript

After that, create the directory /hadoop/disk0 (it doesn’t have to be named like that. you could use a directory of your choice).

JavaScript

Now you are ready to mount the disk:

JavaScript

Finally, you should let hadoop know that you want to use this disk as hadoop storage. Your /etc/hadoop/conf/hdfs-site.xml should contain these config parameters

JavaScript
User contributions licensed under: CC BY-SA
8 People found this is helpful
Advertisement