Back

Explore Courses Blog Tutorials Interview Questions
0 votes
1 view
in Big Data Hadoop & Spark by (11.9k points)

I am trying to learn Hadoop by following a tutorial and trying to do pseudo-distributed mode on my machine.

My core-site.xml is:

<?xml version="1.0"?>

<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>

   <property>

      <name>fs.default.name</name>

      <value>hdfs://localhost:9000</value>

      <description>The name of the default file system. A URI whose scheme and authority determine the FileSystem implementation.       

      </description>   

   </property>

</configuration>

My hdfs-site.xml file is:

<?xml version="1.0"?>

<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>

   <property>

      <name>dfs.replication</name>

      <value>1</value>

      <description>The actual number of replications can be specified when the

        file is created.

      </description>

   </property>

</configuration>

My mapred-site.xml file is:

<?xml version="1.0"?>

<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>

   <property>      

      <name>mapred.job.tracker</name>

      <value>localhost:9001</value>

      <description>The host and port that the MapReduce job tracker runs

        at.

      </description>

   </property>

</configuration>

When I run the command it ran successfully but what it is doing actually:

hadoop-1.2.1$ bin/hadoop namenode -format

14/11/26 12:37:16 INFO namenode.NameNode: STARTUP_MSG: 

/************************************************************

STARTUP_MSG: Starting NameNode

STARTUP_MSG:   host = myhost/127.0.0.8

STARTUP_MSG:   args = [-format]

STARTUP_MSG:   version = 1.2.1

STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2 -r 1503152; compiled by 'mattf' on Mon Jul 22 15:23:09 PDT 2013

STARTUP_MSG:   java = 1.6.0_45

************************************************************/

14/11/26 12:37:17 INFO util.GSet: Computing capacity for map BlocksMap

14/11/26 12:37:17 INFO util.GSet: VM type       = 64-bit

14/11/26 12:37:17 INFO util.GSet: 2.0% max memory = 932118528

14/11/26 12:37:17 INFO util.GSet: capacity      = 2^21 = 2097152 entries

14/11/26 12:37:17 INFO util.GSet: recommended=2097152, actual=2097152

14/11/26 12:37:17 INFO namenode.FSNamesystem: fsOwner=myuser

14/11/26 12:37:17 INFO namenode.FSNamesystem: supergroup=supergroup

14/11/26 12:37:17 INFO namenode.FSNamesystem: isPermissionEnabled=true

14/11/26 12:37:17 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=100

14/11/26 12:37:17 INFO namenode.FSNamesystem: isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)

14/11/26 12:37:17 INFO namenode.FSEditLog: dfs.namenode.edits.toleration.length = 0

14/11/26 12:37:17 INFO namenode.NameNode: Caching file names occuring more than 10 times 

14/11/26 12:37:17 INFO common.Storage: Image file /tmp/hadoop-myuser/dfs/name/current/fsimage of size 115 bytes saved in 0 seconds.

14/11/26 12:37:18 INFO namenode.FSEditLog: closing edit log: position=4, editlog=/tmp/hadoop-myuser/dfs/name/current/edits

14/11/26 12:37:18 INFO namenode.FSEditLog: close success: truncate to 4, editlog=/tmp/hadoop-myuser/dfs/name/current/edits

14/11/26 12:37:18 INFO common.Storage: Storage directory /tmp/hadoop-myuser/dfs/name has been successfully formatted.

14/11/26 12:37:18 INFO namenode.NameNode: SHUTDOWN_MSG: 

/************************************************************

SHUTDOWN_MSG: Shutting down NameNode at chaitanya-OptiPlex-3010/127.0.0.8

************************************************************/

Can someone please let me know what it is doing internally.

How can I check this practically on my machine so I can see the differences before and after running the command. I am new to Hadoop so this can be a trivial question.

1 Answer

0 votes
by (32.1k points)

Hadoop namenode -format

  • Hadoop namenode directory contains the fsimage and edit files which holds the basic information's about hadoop file system such as where is data available, which user created files like that

  • If you format the namenode then the above information's are deleted from namenode directory which is specified in the hdfs-site.xml as dfs.namenode.name.dir

  • But you still have the datas on the hadoop but not namenode meta data

Browse Categories

...