Wednesday, 20 January 2016

Hadoop Training in Chennai

Hadoop Training in Chennai


Hadoop is created by Douglas Reed cutting. Who named hadoop after his child’s stuffed elephant to support Lucene and Nutch search engine products
Open source project administered by Apache software foundation
Hadoop Consists of two key services( HDFS and MedReduce)
Hadoop is a software framework for data intensive computing applications

1. 
Software platform that lets one easily write and run applications that process vast amounts of data. It includes:
– MapReduce – offline computing engine
– HDFS – Hadoop distributed file system
– HBase (pre-alpha) – online data access
2. Yahoo! is the biggest contributor
3. Hadoop implements Google’s MapReduce, using HDFS
4. MapReduce divides applications into many small blocks of work.
5. HDFS creates multiple replicas of data blocks for reliability, placing them on compute nodes around the cluster.
6. MapReduce can then process the data where it is located.
7. Hadoop‘s target is to run on clusters of the order of 10,000-nodes.

Example Applications and Organizations using Hadoop

  • Amazon
  • Yahoo
  • AOL
  • FaceBook
  • FOX interactive media

Why do We Need Hadoop ?

  • Hadoop provides storage for Big Data at reasonable cost
  • Hadoop allows to capture new or more data
  • With Hadoop, you can store data longer
  • Hadoop provides scalable analytics
  • Hadoop provides rich analytics

What qualities/skills in trainees help

  • Good understanding of data warehouse concepts and design patterns
  • Strong experience with Core Java
  • Good experience on Hadoop ,Experience with HDFS, Map-reduce and other tools in Hadoop ecosystem
  • Strong knowledge and hands-on experience with Map-reduce programming model and high level languages like pig or hive
  • Experience with NoSQL data-stores like HBase, Cassandra
  • Understands various configuration parameters and helps arrive at values for optimal cluster performance
  • Knowledge of configuration management / deployment tools like Puppet / Chef
  • Setting up cluster monitoring and alerting mechanism tools like Ganglia, Nagios etc
  • Experience in setting up cross-data center replication
  • Understands how security model using Kerberos and enterprise LDAP product works and helps implement the same

Job Heading

  • Technical Architecture
  • Application Developer
  • Hadoop Developer

No comments:

Post a Comment