click to enable zoom
Loading Maps
We didn't find any results
View Roadmap Satellite Hybrid Terrain My Location Fullscreen Prev Next
Advanced Search

₹ 0 to ₹ 100,000

We found 0 results. Do you want to load the results now ?
Advanced Search

₹ 0 to ₹ 100,000

we found 0 results
Your search results

Spark at Skanda Global Institute

₹ 15,000   ( ₹ 18,000  | 15% Off)
#69/2,2nd Floor,Above Dominos Pizza 8th Cross, BTM 1st Stage, ,
add to favorites

What is Big Data?

Big Data is collection of huge or massive amount of data.We live in data age.And it’s not easy to measure the total volume of data or to manage & process this enormous data. The flood of thisBig Data are coming from different resources.
Such as : New York stock exchange, Facebook, Twitter, AirCraft, Wallmart etc.

Today’s world information is getting doubled after every two years (1.8 times).
And still 80% of data is in unstructured format,which is very difficult to store,process or retrieve. so, we can say all this unstructured data is Big Data.

Why Hadoop is called Future of Information Economy

Hadoop is a Big Data mechanism, which helps to store and process & analysis unstructured data by using any commodity hardware.Hadoop is an open source software framework written in java,which support distributed application.It was introduced by Dough Cutting & Michael J. Cafarellain in mid of 2006.Yahoo is the first commercial user of Hadoop(2008).
Hadoop works on two different generation Hadoop 1.0 & Hadoop 2.0 which, is based on YARN (yet another resource negotatior) architecture.Hadoop named after Dough cutting’s son’s elephant.

Recommended Audience for Bigdata Hadoop Training:

  • IT Engineers and Software Developers
  • Data Warehouse Developers, Java Architects, Data Analysts and SAAS Professionals
  • Students and Professionals aspiring to learn latest technologies and make a career in Big Data using Hadoop.

Pre–Requisites for Hadoop Training:

  • Good analytical skills
  • Some prior experience in Core Java
  • Fundamental knowledge of Unix
  • Basic knowledge of SQL scripting
  • Prior experience in Apache Hadoop is not required

Enroll for expert level Big Data Hadoop Course to build a rewarding career as certified Hadoop developer. Our Hadoop Developer Training course material and tutorials are created by highly experienced instructors. Once you have registered with PrwaTech you will have complete access to our Hadoop video tutorials, course materials, PPT’s, case studies, projects and interview question.

Job Titles for Hadoop Professionals

Job opportunities for talented software engineers in fields of Hadoop and Big Data are enormous and profitable. Zest to become proficient and well versed in Hadoop environment is all that is required for a fresher. Having technical experience and proficiency in fields described below can help you move up the ladder to great heights in the IT industry.


Hadoop Architect

A Hadoop Architect is an individual or team of experts who manage penta bytes of data and provide documentation for Hadoop based environments around the globe. An even more crucial role of a Hadoop Architect is to govern administers, managers and manage the best of their efforts as an administrator. Hadoop Architect also needs to govern Hadoop on large cluster. Every HAdoop Architect must have an impeccable experience in Java, MApreduce, Hive, Hbase and Pig.


Hadoop Developer

Hadoop developer is one who has a strong hold on programming languages such as Core Java,SQL jQuery and other scripting languages. Hadoop Developer has to be proficient in writing well optimized codes to manage huge amounts of data. Working knowledge of Hadoop related technologies such as Hive, Hbase, Flume facilitates him in building an exponentially successful career in IT industry.


Hadoop Scientist

Hadoop Scientist or Data Scientist is a more technical term replacing Business Analyst. They are professionals who generate, evaluate, spread and integrate the humongous knowledge gathered and stored in Hadoop environments. Hadoop Scientists need to have an in-depth knowledge and experience in business and data. Proficiency in programming languages such as R, and tools such as SAS and SPSS is always a plus point.


Hadoop Administrator

With colossal sized database systems to be administered, Hadoop Administrator needs to have a profound understanding of designing principals of HAdooop. An extensive knowledge of hardware systems and a strong hold on interpersonal skills is crucial. Having experience in core technologies such as HAdoop MapReduce,Hive,Linux,Java, Database administration helps him always be a forerunner in his field.


Hadoop Engineer

Data Engineers/ Hadoop Enginners are those can create the data-processing jobs and build the distributed MapReduce algorithms for data analysts to utilize. Data Engineers with experience in Java, and C++ will have an edge over others.


Hadoop Analyst

Big Data Hadoop Analysts need to be well versed in tools such as Impala, Hive, Pig and also a sound understanding of application of business intelligence on a massive scale. Hadoop Analysts need to come up with cost efficient breakthroughs that are faster in jumping between silos and migrating data.

Want to learn the latest trending technology Big Data Hadoop Course? Register yourself for BigData hadoop training classes from the certified bigdata hadoop experts.

Hadoop developer is one who has a strong hold on programming languages such as Core Java,SQL jQuery and other scripting languages. Hadoop Developer has to be proficient in writing well optimized codes to manage huge amounts of data. Working knowledge of Hadoop related technologies such as Hive, Hbase, Flume facilitates him in building an exponentially successful career in IT industry.

Learning Objectives: At the end of Hadoop Developer Training course, participants will be able to:

  • Completely understand Apache Hadoop Framework.
  • Learn to work with HDFS.
  • Discover how MapReduce works with data and processes it.
  • Design and develop big data applications using Hadoop Ecosystem.
  • Learn how YARN helps in managing resources into clusters.
  • Write as well as execute programs in YARN.
  • Implement MapReduce Integration, HBase, Advanced Indexing and Advanced Usage.
  • Work on assignments.

Module 1: Introduction of Big Data & Hadoop Architecture

Topics covered on Basics of Hadoop

In this module, will discuss about Big Data. How Big Data impact in our social life & its important role. How Hadoop is helpful to manage & process Big Data. Hadoop Ecosystem & its Architecture. Hadoop components: HDFS & Mapreduce manage to store & process Big Data.

  • What is Big Data
  • Role of Big Data
  • Hadoop components
  • Hadoop architecture
  • HDFS
  • Map Reduce
  • Name Node
  • Job Tracker
  • Secondary Name Node
  • Data node
  • Data Pipe lining
  • Task Tracker
  • Rack Awareness
  • Anatomy of file read & write

Module 2: Playing around Cluster & set up Hadoop cluster

Topics covered on Hadoop Cluster

In this module, we will learn to set up Hadoop Cluster on five different mode. How to configure important files. Data loading & processing.

  • Hadoop cluster on single Node (pseudo mode)
  • Multiple node cluster
  • Cssh Cluster
  • Configuring Files
  • Data Loading
  • Map reduce processing
  • Recover secondary Name Node
  • Adding & Deleting Data Node
  • Balancing

Module 3: Hadoop MapReduce Framework & impletation

Topics covered on MapReduce Framework

In this module, will work on Map Reduce Framework.How Map Reduce implement on Data which is stored in HDFS . Know about Input split, input format & output format. Overall Map Reduce Process & different stages to process the data.

  • Mapper
  • Reducer
  • Driver
  • Input split
  • Participation
  • Combiner
  • Shuffling
  • Input format
  • Output format
  • Text Input /output format
  • Sequence File format
  • N-Line format
  • Reuse of JVM
  • Record Reader
  • Job scheduler
  • Safe Mode
  • Compression & Decompression (codec, Lzo, snappy)

Module 4: Advanced Map Reduce (Mrv2)

Topics covered on Advanced MapReduce

In this we will work with advanced Map Reduce process complex Data. We are working with new components such as Counters, Distributed Cache to additional data while processing. Custom writable, Serialization.

  • Counters
  • Distributed Cache
  • Data Localization
  • Speculative Execution
  • Reuse of JVM
  • Mrunit Testing
  • Unit testing
  • Advance map reduce framework

Module 5: PIG (Analytics using PIG) & PIG Latin

Topics covered on PIG

In this module, will learn about analytics with PIG. About Pig Latin scripting, complex data type, different cases to work with PIG. Execution environment, operation & transformation.


  • About Pig
  • PIG Installation
  • Pig latin scripting
  • Complex Data Type
  • File Format
  • Where to use PIG when there is MR
  • Operation & transformation
  • Compilation
  • Load, Filter, Join,
  • foreach,
  • Hadoop scripting
  • Pig UDF
  • PIG project

Module 6: Hive & HQL with Analytics

Topics covered on HIVE

In this Module we will discuss a data-ware house package which analysis structure data. About Hive installation and loading data. Storing Data in different Table.

  • About Hive
  • Hive Installation
  • Manage table
  • External table
  • Complex data Type
  • Execution engine
  • Partition & Bucketing
  • Hive UDF
  • Hive querry (sorting , aggregating, Joins, Subquerry)
  • Map reduce side joins
  • Hive project

Module 7: Advance Hive, NoSQL Databases and HBase

Topics covered on Advance Hive, NoSQL and HBase

In this module, you will understand Advance Hive concepts such as UDF. You will also acquire in-depth knowledge of what is HBase, how you can load data into HBase and query data from HBase using client.

  • Hive: Data manipulation with Hive
  • User Defined Functions
  • Appending Data into existing Hive Table
  • Custom Map/Reduce in Hive
  • Hadoop Project: Hive Scripting
  • HBase: Introduction to HBase
  • Client API’s and their features
  • Available Client
  • HBase Architecture
  • MapReduce Integration

Module 8: Advance HBase and ZooKeeper

Topics covered on Advance HBase and Zookeeper

This module will cover Advance HBase concepts. You will also learn what Zookeeper is all about, how it helps in monitoring a cluster, why HBase uses Zookeeper and how to Build Applications with Zookeeper.

  • HBase: Advanced Usage
  • Schema Design
  • Advance Indexing
  • Coprocessors
  • Hadoop Project: HBase tables
  • The ZooKeeper Service: Data Model
  • Operations
  • Implementation
  • Consistency
  • Sessions
  • States

Module 9: Hadoop 2.0, MRv2 and YARN

Topics covered on Hadoop 2.0 and YARN

In this module, you will understand the newly added features in Hadoop 2.0, namely, YARN, MRv2, NameNode High Availability, HDFS Federation, support for Windows etc.

  • Schedulers:Fair and Capacity
  • Hadoop 2.0 New Features: NameNode High Availability
  • HDFS Federation
  • MRv2
  • YARN
  • Running MRv1 in YARN
  • Upgrade your existing MRv1 code to MRv2
  • Programming in YARN framework

Module 10: Hadoop Project Environment

Topics covered on Hadoop Projects

In this module, you will understand how multiple Hadoop ecosystem components work together in a Hadoop implementation to solve Big Data problems. We will discuss multiple data sets and specifications of the project. This module will also cover Apache Oozie Workflow Scheduler for Hadoop Jobs.

Good Course
  • Content
  • Instructor
  • Institute


very good
User Rating 0 (0 votes)
Comments Rating 0 (0 reviews)
spark skanda global bangalore course
Price: ₹ 15,000
Start-End Dates: 22 Oct 16 - 21 Nov 16
Course Duration: 30 days
Discount: 25%
Instructional Level: Appropriate for All
Live Projects
Doubt Clearing Sessions
Reading Material
EMI Option
Online Support
Post completion course access
Practice Exams
Placement assistance
Refund Policy
Post completion support

Compare courses

Leave a Reply

Skanda Global IT Solution

080 – 42096333

Contact Us