Big Data & Hadoop 2.0

Duration: 150 hours training + Project Work + Evaluation Tests

Major Recruiters for Hadoop Developers:

TATA Consultancy Services, Morgan Stanley, WIPRO, Accenture, IBM, NETpro System, Tech Mahindra, Infosys, Cognizant Technology Solutions, HCL Technologies, Capgemini, Peopleclick Techno Solutions, Extrapreneurs India, ATOS, S-Apps Technologies, Quantum Value IT Services, KPISOFT, IQmen Data Solutions, Spiro Solutions, Genpact, Productive Creations India, Human Network, S-Team Software Solutions India, Object Technology Solutions India, Sri Ganesh Hosting, Inventeron Technologies And Business Solutions, National Payments Corporation of India (NPCI), MK Group India, Wild Jasmine, etc.

Designations Avg. Base Salary*
Big Data Engineer INR 746000 per annum
Hadoop Engineer INR 625000 per annum
Hadoop Administrator INR 687000 per annum
Hadoop Developer INR 500000 per annum
Data Engineer INR 454000 per annum
NOTE: There are umpteen types of job opportunities, designations freelancing opportunities with wide variety of payment options other than the ones mentioned in this table and it is practically not possible to list them all. The skills you earn from this program combined with the skills earned by doing projects can catapult your career into unimaginable heights.

*Salary estimates are based on salaries submitted to Glassdoor (April 2020)

Career opportunities:

In this world of ever increasing amount of data, Big Data Analytics has become inevitable for almost all forms of industries ranging from small scale to large scale like Governments, Political Campaigns, Healthcare, IT, Automobile, FMCG, Pharmaceutical, Textile, Media, Education, Insurance, Banking, Internet of Things, Energy, Electronics, Telecommunication, and in fact almost any industry. Big Data developers and analysts are so much in demand in almost all industries that the skill in this area can offer huge opportunities for even entry level employees.
Since 2012, every day 2.5 exabytes of data is estimated to have generated. IDC report 2012 prediction about the growth of global data volume to be 4.4 zettabytes to 44 zettabytes between 2013 and 2020 was pretty accurate. IDC predicts there will be 163 zettabytes of data by 2025. The only way to survive this boom in data is to have more and more Big Data professionals. Facebook alone processes more than 500 terabytes per day. Google receives over 63,000 searches per second on any given day, 3.8 million searches per minute, 228 million searches per hour, and 5.6 billion searches per day, 2 trillion searches per year.

Why learn from us?

Fast Track Program in Cloud Computing (FTP CC) is a tailormade job-oriented training program for Indian students. It was developed by a dedicated team of professionals and academicians over a period not less than 4 years of extensive and intensive research and practice. Institutes and online learning platforms that provide our training programs maintain very limited intakes to individual batches to ensure that Student: Faculty ratio is optimal for efficient delivery of classes.

  • Learning Type: Strictly LIVE Instructor-led and NOT Pre-Recorded Video Tutorials/WhatsApp Training. This ensures that students get to clear their doubts effectively and move forward to each chapter/session/module with clear understanding of the topics covered.
  • Instructors: The instructors at all NGJ Member Institutes are provided by NGJ and hence the quality of training shall be top-notch.
  • Cost: Determined by respective NGJ Member Institutes
  • Certification Test: The evaluation process of the certification test is transparent and efficiently conducted directly by the Board of Examination of NGJ and not by any of the NGJ Member Institutes.
Program Framework
Module 1 – LINUX

Access the command line

  • Log in to a Linux system and run simple commands using the shell
  • Hands On

Manage files from the command line

  • Copy, move, create, delete, and organize files from the bash shell prompt
  • Hands On

Get help in Red Hat Enterprise Linux

  • Resolve problems by using online help systems and Red Hat support utilities.
  • Hands On

Create, view, and edit text files

  • Create, view, and edit text files from command output or in an editor.
  • Hands On

Manage local Linux users and groups

  • Manage local Linux users and groups and administer local password policies.
  • Control access to files with Linux file system permissions Set Linux file system permissions on files and interpret the security effects of different permission settings.
  • Hands On

Monitor and manage Linux processes

  • Obtain information about the system, and control processes running on it.
  • Hands On

Control services and Daemons

  • Control and monitor network services and system daemons using systemd
  • Hands On

Configure and secure OpenSSH service

  • Access and provide access to the command line on remote systems securely using OpenSSH
  • Hands On

Analyze and store logs

  • Locate and accurately interpret relevant system log files for troubleshooting purposes.
  • Hands On

Manage Red Hat Enterprise Linux networking

  • Configure basic IPv4 networking on Red Hat Enterprise Linux systems.
  • Hands On

Archive and copy files between systems

  • Archive files and copy them from one system to another
  • Hands On

Install and update software packages

  • Download, install, update, and manage software packages from Red Hat and yum package repositories.
  • Hands On

Access Linux file systems

  • Access and inspect existing file systems on a Red Hat Enterprise Linux system.
  • Hands On

Use Regular Expressions with Grep

  • Write regular expressions that, when partnered with grep, will allow you to quickly isolate or locate content within text files.
  • Hands On

Create and Edit text files with VIM

  • Introduce the vim text editor, with which you can open, edit, and save text files.
  • Hands On

Comprehensive Module 1 review: Practice and demonstrate the knowledge and skills learned.

Module 2 - JAVA

Introduction

  • Programming language Types and Paradigms
  • Computer Programming Hierarchy
  • How Computer Architecture Affects a Language?
  • Why Java?
  • Flavors of Java
  • Java Designing Goal
  • Role of Java Programmer in Industry
  • Features of Java Language.
  • JVM –The heart of Java
  • Java’s Magic Bytecode

Language Fundamentals

  • The Java Environment:
  • Installing Java.
  • Java Program Development
  • Java Source File Structure
  • Compilation
  • Executions.

Object Oriented Programming

  • Class Fundamentals.
  • Object & Object reference.
  • Object Lifetime & Garbage Collection
  • Creating and Operating Objects
  • Constructor & initialization code block
  • Access Control, Modifiers, methods
  • Nested, Inner Class & Anonymous Classes
  • Abstract Class & Interfaces
  • Defining Methods, Argument Passing Mechanism
  • Method Overloading, Recursion
  • Dealing with Static Members and Finalize () Method
  • Native Method. Use of “this” reference.
  • Use of Modifiers with Classes & Methods
  • Design of Accessors and Mutator Methods
  • Cloning Objects, shallow and deep cloning
  • Generic Class Types

Extending Classes and Inheritance

  • Use and Benefits of Inheritance in OOP
  • Types of Inheritance in Java
  • Inheriting Data Members and Methods
  • Role of Constructors in inheritance
  • Overriding Super Class Methods.
  • Use of “super”
  • Polymorphism in inheritance
  • Type Compatibility and Conversion
  • Implementing interfaces

Package

  • Organizing Classes and Interfaces in Packages
  • Package as Access Protection
  • Defining Package
  • CLASSPATH Setting for Packages
  • Making JAR Files for Library Packages
  • Import and Static Import
  • Naming Convention for Packages

Exception Handling

  • The Idea behind Exception
  • Exceptions & Errors
  • Types of Exception
  • Control Flow in Exceptions
  • JVM reaction to Exceptions
  • Use of try, catch, finally, throw, throws in Exception Handling
  • In-built and User Defined Exceptions
  • Checked and Un-Checked Exceptions

Array & String

  • Operation on String
  • Mutable & Immutable String
  • Using Collection Bases Loop for String
  • Tokenizing a String
  • Creating Strings using StringBuffer

Thread

  • Thread Life Cycle
  • Thread Priorities
  • Synchronizing Threads
  • Inter Communication of Threads
  • Critical Factor in Thread Deadlock

Applet

  • Applet & Application
  • Applet Architecture
  • Parameters to Applet
  • Embedding Applets in Web page
  • Applet Security Policies

A Collection of Useful Classes

  • Utility Methods for Arrays
  • Observable and Observer Objects
  • Date & Times
  • Using Scanner
  • Input/output Operation in Java (java.io Package)
  • Streams and the new I/O Capabilities
  • Understanding Streams
  • The Classes for Input and Output
  • The Standard Streams
  • Working with File Object
  • Reading and Writing to Files
  • Buffer and Buffer Management
  • Read/Write Operations with File Channel
  • Serializing Objects

GUI Programming

  • Designing Graphical User Interfaces in Java
  • Components and Containers
  • Basics of Components
  • Using Containers
  • Layout Managers
  • AWT Components
  • Adding a Menu to Window
  • Extending GUI Features Using Swing Components
  • Java Utilities (java.util Package)

The Collection Framework

  • Collections of Objects
  • Collection Types
  • Sets
  • Sequence
  • Map
  • Understanding Hashing
  • Use of ArrayList & Vector

Event Handling

  • Event-Driven Programming in Java
  • Event- Handling Process
  • Event-Handling Mechanism
  • The Delegation Model of Event Handling
  • Event Classes
  • Event Sources
  • Event Listeners
  • Adapter Classes as Helper Classes in Event Handling
  • Anonymous Inner classes a Short cut to Event Handling
  • Avoiding Deadlocks in GUI Code
  • Event Types & Classes
  • Networking Programming
  • Networking Basics
  • Client-Server Architecture
  • Socket Overview
  • Networking Classes and Interfaces
  • Network Protocols
  • Developing Networking Applications in Java

Database Programming using JDBC

  • Introduction to JDBC
  • JDBC Drivers & Architecture
  • CURD operation Using JDBC
  • Connecting to non-conventional Databases
  • Role of JSP in MVC-2
  • JSP with Database
  • JSP Implicit Objects
  • Tag Libraries
  • JSP Expression Language (EL)
  • Using Custom Tag
  • JSP Capabilities:
  • Exception Handling
  • Session Management
  • Directives
  • JSP with Java Bean

Comprehensive Module 2 review: Practice and demonstrate the knowledge and skills learned.

Module 3 - Big Data & Hadoop 2.0

Introduction to Big Data & Hadoop

  • What is Big Data
  • Limitations of existing models on processing Big Data
  • Apache Hadoop
  • Hadoop architecture
  • Hadoop Storage: HDFS
  • The framework of Map Reduce
  • Practice time: This will be mostly introductory session, and students will be given overview about the course and technology
  • A look at the interview questions asked

Getting Started with Hadoop

  • Understanding system requirements for installing Hadoop and VMware player
  • Virtualization and the VMware player
  • Installing the VMware player
  • Creating a virtual machine in the VMware player
  • Practice time: Students will Download and install VMware player on their machines and create a virtual machine instance on the installed software.

Hadoop Architecture

  • Architecture Overview
  • Hadoop Configuration
  • Hadoop Core Services
  • Hadoop Storage – HDFS
  • Hadoop Storage – HDFS
  • Regular file System Vs HDFS
  • Data Block Split
  • HDFS – Block Replication Architecture
  • Practice time: Students will be shown a small demo of Hadoop in action, various services running, different commands in HDFS and map reduce. This will be done to give them a better understanding of the big picture and how things can be expected to work for big data in Hadoop architecture.

Hadoop Deployment

  • Hadoop Installation pre-requisite
  • Hadoop Single Node cluster Installation
  • Hadoop Multi Node Cluster Installation
  • Single Node Cluster Vs Multi Node Cluster
  • Practice time: Students will be installing Hadoop in their machine and creating single as well as multi-node cluster. A sample job will be given to students to run on both types of cluster setups

Introduction to Map Reduce

  • What is Map Reduce
  • Map Execution
  • Map Reduce Essentials
  • Map Reduce Jobs
  • Map Reduce Engine
  • Checking Hadoop environment for Map Reduce
  • Taking sample Big Data and writing map reduce code for it
  • Practice time: Students will be given different problem statements and data sets and asked to write map-reduce code on those data sets, executing and checking the output on their single node clusters.

Introduction to Pig

  • What is Pig
  • How it works
  • Data Model
  • Pig Execution and Interactive modes
  • Pig Vs SQL
  • Installing a Pig Engine
  • Various Operations
  • Pig Commands
  • Practice time: Students will be practicing various Pig commands trying to query different formats of data from various sample data provided to them

Introduction to Hive

  • What is Hive and why was it developed
  • Various components of Hive
  • Hive Data Model
  • Writing simple queries in Hive
  • Comparing Hive Queries with corresponding PIG commands
  • Practice time: In this hands-on session students will execute various Hive queries on different types of sample data provided to them by the trainer

Comprehensive Module 3 review: Practice and demonstrate the knowledge and skills learned.

Final Certification Test: Students will have to give a test after the session to determine how much they have learned; this test will be objective as well as subjective where students have to write code and execute it in front of the trainer to get the grades.

Follow up Project: Students are given follow-up projects and large data sets on which they can write Map Reduce code and process those on their cluster. This will help them with deeper understanding of Hadoop for solving complicated problems.