Intellipaat Back

Explore Courses Blog Tutorials Interview Questions
0 votes
in Big Data Hadoop & Spark by (55.6k points)

a. 1
b. 2
c. 3
d. 4

1 Answer

0 votes
by (119k points)

The correct answer is option A (one). Job Tracker is a daemon service that helps in submitting and tracking MapReduce jobs in Hadoop. There is only one instance of a job tracker that can run on Hadoop Cluster. Job tracker can be run on the same machine running the Name Node but in a typical production cluster its run on a separate machine.

In case you want to learn Hadoop, I suggest enrolling for this Big Data course by Intellipaat.

You can watch this video on Hadoop by Intellipaat to learn installing Hadoop and get started with Hadoop:

Browse Categories