Hadoop online training classes(1)

Page 1

Newyorksys.com



Contents          

What is Hadoop. Why do we need Hadoop. How Hadoop works. HDFS Architecture. What is Map – Reduce. Hadoop Cluster. Hadoop Processes. Topology of a Hadoop Cluster. Distinction of Hadoop Framework . Prerequisites to learn hadoop.


What is Hadoop  Hadoop is an open Source Framework.  Developed by Apache Software Foundation.

 Used for distributed processing of large date

sets.  It works across clusters of computers using a simple programming model (Map-Reduce).


Why do we need Hadoop  Data is growing faster.  Need to process multi petabytes of data.  The performance of traditional applications

is decreasing.  The number of machines in a cluster is not constant.  Failure is expected, rather than exceptional.


How Hadoop Works  The Hadoop core consists of two modules :  Hadoop Distributed File System (HDFS)

[Storage].  Map Reduce [Processing]. Mapper Reducer


HDFS Architecture


What is Map – Reduce  Map Reduce plays a key role in hadoop framework.  Map Reduce is a Programming model for writing

applications that rapidly process large amount of data.  Mapper – is a function that processes input data to generate intermediate output data.  Reducer – Merges all intermediate data from all mappers and generate final output data.


Hadoop Cluster  A Hadoop Cluster consist of multiple

machines Which can be classified into 3 types   

Namenode Secondary Namenode Datanode


Hadoop Processes ď‚— Below are the daemons (Processes) Which

runs in a cluster.

Name node (Runs on a master machine) Job Tracker (Runs on a master machine) Data node (Runs on slave machines) Task Tracker (Runs on slave machines)


Topology of a Hadoop Cluster


Distinction  Simple – Hadoop allows users to quickly

write efficient parallel code.  Reliable – Because Hadoop runs on commodity hardware, it can face frequent automatically handle such failures.  Scalable – we can increase or decrease the number of nodes (machine) in hadoop cluster.


Prerequisites  Linux bases operating system (Mac OS, Redhat,      

ubuntu) Java 1.6 or higher version Disk space ( To hold HDFS data and it’s replications ) Ram (Recommended 2GB) A cluster of computers. You can even install Hadoop on single machine. For more information visit this link: hadoop online training


Newyorksys.com ď‚— NewyorkSys is one of

the leading Training Company in US. We have certified trainers. We provide Online Training, Fast Track online training, with job assistance. We are providing excellent Training in all courses. We also help you in resume preparation and provide job assistance till you get job. For more details Visit : http://www.newyorksys.com 15 Roaring Brook Rd, Chappaqua, NY 10514. USA: +1-718-313-0499 , 718-305-1757 Email : enquiry@newyorksys.us


About Newyorksys Newyorksys , the leading IT online training company is going to conduct its exclusive Hadoop online training programs. We have a comprehensive and engaging curriculum for delivering the hadoop concepts Labels: hadoop online training, hadoop training, hadoop course, hadoop online course, hadoop training in usa, hadoop technologies, hadoop training online, hadoop classes in usa, hadoop training and development, online hadoop training, apache hadoop training, apache hadoop course in usa


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.