Learners 1000+
  • About Hadoop -

    Hadoop is a Big Data technology that provides various tools to work with Big Data. Big Data means in general data that starts from tera bytes and peta bytes means very huge data like for banking, call center, insurance, google, facebook etc.

  • About Hadoop / Big Data Certification Course -

    This course can be learnt by both Freshers and Experienced who wish to become a Hadoop Big Data Expert. Learn from basics to adv 10+ techniques by Hadoop to work with Big Data.

  • What You Will Learn ?

    Installation on Linux, Scoop, Flume, Map Reduce etc. 10+ Tools

  • Trainer -

    Hadoop Expert

  • Duration -

    1.5 to 2 months (35 hours)

  • Any Pre-requisites -

    None. Any one from IT & non IT can learn.(No programming)

  • Scope & Job Openings -

    RLimited openings and Comes under Hot Skills hence High Pay.

Syllabus

30 hours
  • HDFS - Hadoop Distributed File System
  • Comparing Hadoop & SQL
  • Hadoop Architecture
  • Map Reduce & HDFS
  • Using the Hadoop single node image (Clone)
  • HDFS Design & Concepts
  • Blocks, Name nodes and Data nodes
  • HDFS High-Availability and HDFS Federation
  • Hadoop DFS The Command-Line Interface
  • Basic File System Operations
  • Anatomy of File Read & Write
  • Block Placement Policy and Modes
  • More detailed explanation about Configuration files
  • Metadata, FS image, Edit log, Secondary Name Node and Safe Mode
  • FSCK Utility. (Block report)
  • HDFS Federation
  • Introduction
  • Import Data
  • Export data
  • Introduction
  • Flume Commands
  • Functional Programming Basics
  • Map and Reduce Basics
  • Anatomy of a Map Reduce Job Run
  • Architecture
  • Job Completion, Failures
  • Shuffling and Sorting
  • Splits, Record reader, Partition, Types of partitions & Combiner
  • Speculative Execution
  • Types of Schedulers and Counters
  • YARN
  • Sequential Files and Map Files
  • Hands on “Word Count” in Map/Reduce in standalone and Pseudo distribution Mode
  • Installation
  • Introduction and Architecture
  • Hive Services, Hive Shell, Hive Server and Hive Web Interface (HWI)
  • Meta store
  • Hive QL
  • Derby Database
  • Working with Tables
  • Primitive data types and complex data types
  • Working with Partitions
  • Hive Bucketed Tables and Sampling
  • External partitioned tables
  • Differences between ORDER BY, DISTRIBUTE BY and SORT BY
  • Log Analysis on Hive
  • Hands on Exercises
  • Introduction
  • Types
  • Brief Of Nosql Database
  • HBase Installation
  • HBase concepts
  • HBase Data Model and Comparison between RDBMS and NOSQL
  • Master & Region Servers
  • HBase Operations (DDL and DML) through Shell and Programming and HBase Architecture
  • Catalog Tables
  • Block Cache and sharing
  • SPLITS
  • DATA Modeling
  • HBASE Filters
  • Bulk Loading and Coprocessors
  • Real world use case consisting of HDFS, MR and HBASE
  • Spark and its purpose
  • Components of Spark
  • Using Spark with Hadoop
  • Resilient Distributed Dataset (RDD)
  • Working with Key-Value Pairs
  • Downloading and installing Spark Standalone
  • Scala Overview
  • Fundamental Of Scala
  • Basic Syntax and Operators
  • Programming with Spark
  • Fundamental Of Solr
  • Installation of Solr Standalone
  • Basic Solr Concept
  • Demo

Why get Trained from us ?

10 yrs in IT Training
ISO Certified Institute
Training from Experts
Trained Globally
Certification Course
100% Practical Training
Reasonable Fees
Interview Preparation
Live Projects / Case Studies
Free Demo
1000+ Companies
Best Placement Service
Lots of students got placed in ongoing training or within just 30 days of completion of training!

You can be the Next Placed Student!