Horje
Install Hadoop on Mac

Apache Hadoop is a strong framework based on open sources that is capable of implementation of distributed storage and processing of massive data volumes across a system made up of a network of computers. It is the favoured technology for the big data steel Industry, that is, its scalability, reliability, and fault-protecting characteristics are superior. Let’s take you through the Hadoop installation process on a Mac OS computer using this guide.

Installation Prerequisites

Before we proceed, ensure that the necessary conditions below are met on your macOS system:

1. Java Development Kit (JDK): Hadoop is written in Java and therefore needs it to function. Ensure that you have the JDK 8 or higher packed and ready for installation on your computer. You can get JDK either using a website Oracle (as an option) or a HomeBrew project, as brew installs OpenJDK.

2. SSH: Hadoop SSH communication is tied to Hadoop’s ability to transmit data between different nodes in the cluster. macOS usually comes with SSH installed by default, however, to enable the feature you will need to navigate to System Preferences -> Sharing -> Remote Login.

3. Homebrew: Homebrew, the package manager for macOS, lists software components in their preferred software location. With Homebrew not being installed, use the following command on your shell.

a20

Homebrew Package

4. the Environment Setup: Make sure that the environment variable is correctly set up. This responsibility can be taken up by editing the ~/.bash_profile or ~/.zshrc file and adding in the mentioned lines.

a21

Setting Up The Environment

  • Run the replace /path/to/your/java/home and replace /path/to/your/Hadoopwith/home with your full path to the installed Java and Hadoop.

How to Install Hadoop on Mac?

Apache Hadoop is a strong framework based on open sources and we need to follow the below-mentioned easy steps to download it properly on MacOS.

Step 1: Downloading Hadoop

You can download the latest stable version of Hadoop from the Apache Hadoop website. Select the one that fits your requirements and download the tarball (tar.gz) from the server to your machine.

Then, download the files, then extract the file into any of the directories that you wish.

For example

a22

Extracting Tarball files

Step 2: Configuration

Navigate to the Hadoop configuration directory (in $HADOOP_HOME/etc/Hadoop) and modify the given configuration as indicated below.

a. core-site.xml:

a23

Hadoop Configuration

b. hdfs-site.xml:

a24

Hadoop Configuration

c. mapred-site.xml (if it doesn’t exist, create it):

a25

mapred-site.xml

d. yarn-site.xml:

a26

yarn-site.xml

Step 3: Formatting HDFS

The prerequisite for starting Hadoop services is formatting the disk HDFS (Hadoop Distributed File System) by the end of it.

Run the following command

a27

Formatting HDFS

Step 4: Start Hadoop Services.

Launch Hadoop services by executing the commands mentioned below.

a. Start HDFS:

a28

Starting HDFS

b. Start YARN:

a29

Start YARN

Step 5: Verify Installation

Hadoop is a program that you can check its correct running by viewing the Hadoop web interface. Open your web browser and go to http://localhost:9876 for the HDFS NameNode interface and http://localhost:6060 for the YARN ResourceManager Interface.

a30

Output of Hadoop running on MacOs

Conclusion

Congratulations! You have successfully installed Apache Hadoop on macOS. You can now move on to the full spectrum of possibilities afforded by Hadoop and start investigating its role in distributed data processing and analysis. Keep in mind that the official Hadoop documentation can be of help to those interested in more advanced configurations and more complex deployments.

Also Read

Install Hadoop on Mac – FAQs

Can I install Hadoop on macOS without using Homebrew?

Yes, you would be able to download and set up Hadoop on macOS without getting Access to Homebrew. Hence, Homebrew provides users with a simplified process that should be done automatically by the handling of the dependencies and updates.

Do I need to set up a multi-node cluster for Hadoop on macOS?

No, you can set up and run Hadoop on one Mac OSX machine for development and testing. For multiple machines configuration you need a multi-node cluster setup as it includes necessary network settings too.

What is the purpose of formatting HDFS during Hadoop installation?

Formatting HDFS creates the underlying storage units and the directories automatically. The application will seek layers of security to gain access. It acts as such a step in order to keep the HDFS work well and effectively avoid data corruption.

How can I change the replication factor for HDFS on macOS?

The number of copies that are created when a file is stored in the Hadoop file system can be changed by correcting the value of the ‘dfs.replication’ property in ‘hdfs-site.xml’ configuration file. The initial replication level has a value of 3 (by default), but according to your data storage and reliability needs, you can change it.

Is it necessary to set up environment variables for Hadoop on macOS?

Its strongly recommended to set up environment variables such as ‘JAVA_HOME ‘and ‘HADOOP_HOME’, especially, it will prevent you from using Hadoop commands in different terminal windows all over again. This way it helps with managing Hadoop install path and Java space.




Reffered: https://www.geeksforgeeks.org


Installation Guide

Related
How to Install and Uninstall Software via CMD? How to Install and Uninstall Software via CMD?
How to Install Docker on Debian? How to Install Docker on Debian?
How to Bypass "Lets connect you to a network" in Windows 11? How to Bypass "Lets connect you to a network" in Windows 11?
How to extract .deb file in Ubuntu? How to extract .deb file in Ubuntu?
How to install Ruby Programming Language on Manjaro Linux? How to install Ruby Programming Language on Manjaro Linux?

Type:
Geek
Category:
Coding
Sub Category:
Tutorial
Uploaded by:
Admin
Views:
16