Pages

Thursday 22 April 2021

Hadoop-Ecosystem

Introduction: Hadoop Ecosystem is a platform or a suite which provides various services to solve the big data problems. It includes Apache projects and various commercial tools and solutions. There are four major elements of Hadoop i.e. HDFS, MapReduce, YARN, and Hadoop Common. Most of the tools or solutions are used to supplement or support these major elements. All these tools work collectively to provide services such as absorption, analysis, storage and maintenance of data etc.

Following are the components that collectively form a Hadoop ecosystem:

HDFS: Hadoop Distributed File System
YARN ,Map Reduce: Yet Another Resource Negotiator,: Programming based Data Processing
Sqoop: Data exchange
Flume: Log collector
Hive: Query based processing of data services
Pig: Query based processing of data services
Mahout: Machine Learning algorithm libraries
R Connectors: Statistical data 
Ambari: Managing and monitoring Hadoop clusters
Zookeeper: Managing cluster
Oozie: Job Scheduling
Hbase: NoSQL Database


All these toolkits or components revolve around one term i.e. Data. That’s the beauty of Hadoop that it revolves around data and hence making its synthesis easier.

HDFS:
HDFS is the primary or major component of Hadoop ecosystem and is responsible for storing large data sets of structured or unstructured data across various nodes and thereby maintaining the metadata in the form of log files.
HDFS consists of two core components i.e.
Name node
Data Node
Name Node is the primary node which contains metadata (data about data) and the data nodes that stores the actual data. These data nodes are commodity hardware in the distributed environment.


YARN ,MapReduce
Yet Another Resource Negotiator, as the name implies, YARN is the one who helps to manage the resources across the clusters. In short, it performs scheduling and resource allocation for the Hadoop System.
Consists of three major components i.e.
Resource Manager
Nodes Manager
Application Manager













MapReduce:
MapReduce makes it possible to carry over the processing’s logic and helps to write applications which transform big data sets into a manageable one. MapReduce makes the use of two functions i.e. Map() and Reduce() whose task is:
Map() performs sorting and filtering of data and thereby organizing them in the form of group. Map generates a key-value pair based result which is later on processed by the Reduce() method.
Reduce(), as the name suggests does the summarization by aggregating the mapped data. In simple, Reduce() takes the output generated by Map() as input and combines those tuples into smaller set of tuples.









 
Sqoop:
Sqoop is a tool designed to transfer data between Hadoop and relational database servers. It is used to import data from relational databases (such as Oracle and MySQL) to HDFS and export data from HDFS to relational databases.












Flume:
Apache Flume is a reliable and distributed system for collecting, aggregating and moving massive quantities of log data. Apache Flume is used to collect log data present in log files from web servers and aggregating it into HDFS for analysis.











Hive:
Apache Hive, is an open source data warehouse system for querying and analyzing large datasets stored in Hadoop files. Hive do three main functions: data summarization, query, and analysis. Hive use language called HiveQL (HQL), which is similar to SQL.












PIG:
Pig was basically developed by Yahoo which works on a pig Latin language, which is Query based language similar to SQL.
It is a platform for structuring the data flow, processing and analyzing huge data sets.
Pig does the work of executing commands and in the background, all the activities of MapReduce are taken care of. After the processing, pig stores the result in HDFS.











Mahout:
Apache Mahout is an open source project that is primarily used for creating scalable machine learning algorithms.
It provides various libraries or functionalities such as collaborative filtering, clustering, and classification which are nothing but concepts of Machine learning. It allows invoking algorithms as per our need with the help of its own libraries.

R Connectors:
Oracle R Connector for Hadoop provides API access from a local R client to Hadoop, using these APIs: 
hadoop : Provides an interface to Hadoop MapReduce. 
hdfs : Provides an interface to HDFS. 
orhc : Provides an interface between the local R instance and Oracle Database.

Ambari:
Ambari, another Hadop ecosystem component, is a management platform for provisioning, managing, monitoring and securing apache Hadoop cluster. Hadoop management gets simpler as Ambari provide consistent, secure platform for operational control.











Zookeeper:
There was a huge issue of management of coordination and synchronization among the resources or the components of Hadoop which resulted in inconsistency, often. Zookeeper overcame all the problems by performing synchronization, inter-component based communication, grouping, and maintenance.











Oozie:
Oozie simply performs the task of a scheduler, thus scheduling jobs and binding them together as a single unit. There is two kinds of jobs .i.e Oozie workflow and Oozie coordinator jobs. 
Oozie workflow is the jobs that need to be executed in a sequentially ordered manner areas 
Oozie Coordinator jobs are those that are triggered when some data or external stimulus is given to it.











Hbase:
Apache HBase is a Hadoop ecosystem component which is a distributed database that was designed to store structured data in tables that could have billions of row and millions of columns. HBase is scalable, distributed, and NoSQL database that is built on top of HDFS. HBase, provide real-time access to read or write data in HDFS.





HADOOP VENDORS:

No comments:

Post a Comment

Friends-of-friends-Map Reduce program

Program to illustrate FOF Map Reduce: import java.io.IOException; import java.util.*; import org.apache.hadoop.conf.Configuration; import or...