Setting-up a single Hadoop node needs the administrator to abide by certain essential criteria:
||Hadoop Single Node
||Windows or Linux
||Non-distributed mode as a single Java process
||Testing, prototyping, future scaling of nodes
One the above criteria are met, following steps need to be carried out:
Step 1: Download Hadoop
For Hadoop to be installed there are some certain pre-requisites such as Supportive platforms like GNU/Linux or Win32 and softwares like Jawed and SSH must be installed and enabled; a proper environment is needed for the system to be installed. If your Hadoop cluster does not come with all the requisites then you will have to download and install them by yourself which is not that difficult. You can download the latest Jaw version compatible with Hadoop from following link :
Further Configuration of SSH access is needed to enable master/principal and secondary node to get access to the system and take over the slave nodes and handle local users and machines. Generate an SSH key, log in and run the command of key generation. The last step of this process is to check whether machine/machines are connected and working with the main user enabling the local host to permanently combine with other known hosts.
Step 2: Hadoop Installation
You will need a Hadoop distribution which is available on Apache Download Mirrors:
To make things lot easier you can install softwares like Ubunto-64 bit or MVPlayer
In Ubuntu you have an option of Downloading Hadoop by its latest release, choose to download a stable release, unpack the download and extract it.
Step 3: Configure jaw path
This will start the Hadoop installation process and the next step of the installation will be configuration of your Jaw path. The step is simple as you just have to edit the pre defined perimeter of the path in the system and change it according to your user defined path.
Learn the core concepts of Hadoop with this Hadoop tutorial!
Step 4: Edit the location
Edit the built in changeable location address according to your own Jaw address and save it.
Step 5: Add Hadoop System User
In the next step would be to add dedicated Hadoop system user to make the system operational. This step isn’t that much important but it will help keeping Hadoop segregated from other applications, user IDs and software.
Now disable IPv6 as it is necessary to unbind Hadoop but if you are connected to IPv6 network then there is no need to disable it.
Step 6: Format and Edit the Hadoop nodes
The last step is to format and edit the nodes of Hadoop file system, which is only necessary if the system is not already in use. The process is performed by a simple command
Now you can make sure whether the software is fully operational or not by giving a task to the system, hopefully this will work and you will have a smile on your face.
Hadoop can be stopped instantly by commands given below;
Learn what skills you need to become a successful Big Data professionals with these Top Hadoop Interview Questions!
Hadoop has many web interfaces which are very much user friendly, available at these addresses;
This ends the whole process and now you can enjoy the ease of handling your big data conveniently and effectively.
However working on Hadoop system will require a professional to get Hadoop training which will improve his expertise and skill. As Big data technologies are very imperative to use and are very helpful managing the huge volumes of data. Intellipaat lets you master this popular technology in a most efficient way by emphasizing practical implementation of Big Data Hadoop in numerous ways.
Get enrolled in Big Data Hadoop Online Training Course and soar in your career!