Newyorksys.com
Contents
What is Hadoop. Why do we need Hadoop. How Hadoop works. HDFS Architecture. What is Map – Reduce. Hadoop Cluster. Hadoop Processes. Topology of a Hadoop Cluster. Distinction of Hadoop Framework . Prerequisites to learn hadoop.
What is Hadoop Hadoop is an open Source Framework. Developed by Apache Software Foundation.
Used for distributed processing of large date
sets. It works across clusters of computers using a simple programming model (Map-Reduce).
Why do we need Hadoop Data is growing faster. Need to process multi petabytes of data. The performance of traditional applications
is decreasing. The number of machines in a cluster is not constant. Failure is expected, rather than exceptional.
How Hadoop Works The Hadoop core consists of two modules : Hadoop Distributed File System (HDFS)
[Storage]. Map Reduce [Processing]. Mapper Reducer
HDFS Architecture
What is Map – Reduce Map Reduce plays a key role in hadoop framework. Map Reduce is a Programming model for writing
applications that rapidly process large amount of data. Mapper – is a function that processes input data to generate intermediate output data. Reducer – Merges all intermediate data from all mappers and generate final output data.
Hadoop Cluster A Hadoop Cluster consist of multiple
machines Which can be classified into 3 types
Namenode Secondary Namenode Datanode
Hadoop Processes ď‚— Below are the daemons (Processes) Which
runs in a cluster.
Name node (Runs on a master machine) Job Tracker (Runs on a master machine) Data node (Runs on slave machines) Task Tracker (Runs on slave machines)
Topology of a Hadoop Cluster
Distinction Simple – Hadoop allows users to quickly
write efficient parallel code. Reliable – Because Hadoop runs on commodity hardware, it can face frequent automatically handle such failures. Scalable – we can increase or decrease the number of nodes (machine) in hadoop cluster.
Prerequisites Linux bases operating system (Mac OS, Redhat,
ubuntu) Java 1.6 or higher version Disk space ( To hold HDFS data and it’s replications ) Ram (Recommended 2GB) A cluster of computers. You can even install Hadoop on single machine. For more information visit this link: hadoop online training
Newyorksys.com ď‚— NewyorkSys is one of
the leading Training Company in US. We have certified trainers. We provide Online Training, Fast Track online training, with job assistance. We are providing excellent Training in all courses. We also help you in resume preparation and provide job assistance till you get job. For more details Visit : http://www.newyorksys.com 15 Roaring Brook Rd, Chappaqua, NY 10514. USA: +1-718-313-0499 , 718-305-1757 Email : enquiry@newyorksys.us
About Newyorksys Newyorksys , the leading IT online training company is going to conduct its exclusive Hadoop online training programs. We have a comprehensive and engaging curriculum for delivering the hadoop concepts Labels: hadoop online training, hadoop training, hadoop course, hadoop online course, hadoop training in usa, hadoop technologies, hadoop training online, hadoop classes in usa, hadoop training and development, online hadoop training, apache hadoop training, apache hadoop course in usa