Big Data Internship Program - Data Ingestion

Raju Shreewastava & R., All About Big Data & Hadoop

Play Speed
  • 0.5x
  • 1x (Normal)
  • 1.25x
  • 1.5x
  • 2x
29 Videos (2h 20m)
    • Introduction to Data Ingestion

      5:53
    • Recap - Big data Internship Program - Part 1 Foundation

      3:52
    • Data Ingestion Tools

      4:48
    • Some more Data Ingestion Tools

      5:38
    • Introduction to FileFormats

      4:05
    • Text/CSV file formats

      3:01
    • BinaryFileFormats-Sequence Files

      2:48
    • BinaryFileFormats-Avro

      4:24
    • Columnar formats-RC and ORC files

      4:31
    • Columnar format-Parquet Files

      4:02
    • Introduction to sqoop

      5:53
    • Sqoop Import

      4:24
    • Import data from MySql to HDFS

      6:38
    • Other variations of Sqoop Import Command

      5:33
    • Running a Sqoop Export Command

      5:50
    • Sqoop Jobs

      5:41
    • Sqoop incremental import

      5:46
    • Lab: Sqoop incremental Import

      5:30
    • What is Flume?

      2:31
    • Data Flow Model

      4:24
    • Flume Configuration File

      4:25
    • HelloWorld example in Flume

      6:09
    • Multi Agent flow

      2:38
    • Multiplexing

      5:05
    • Interceptors in Flume

      2:30
    • Book recommendation Project Overview

      3:04
    • Book recommendation Project Sqoop Work Part-1

      11:27
    • BookReccomendation Project- Sqoop Work -Part2

      2:38
    • Book recommendation Project - Flume Work

      7:05

About This Class

This course is a part of “Big data Internship Program”  which is aligned to a typical Big data project life cycle stage.

  • Foundation
  • Ingestion
  • Storage
  • Processing
  • Visualization

This course is focused on the Ingestion in Big data . 

Our Course is divided into two part 1) Technical Knowledge with examples and 2) Work on project 

Technical Knowledge 

  1. Big Data ingestion's concept and means 
  2. Sqoop concept and feature.  
  3. Good understanding of sqoop tools with arguments 
  4. Flume concept and configuration 
  5. Flume features: Multiplexing,Flume Agents,Interceptors etc .
  6. Understanding of different File Format supported by Hadoop 

Project Part 

  1. Get the access to our private GitHub repository 
  2. Build the first part of Our Recommendation Book project using sqoop and flume

10

Students

--

Projects

0

Reviews (0)

Raju Shreewastava & R.

All About Big Data & Hadoop

Big Data Trunk is the leading Big Data focus consulting and training firm founded by industry veterans in data domain. It helps is customer gain competitive advantage from open source, big data, cloud and advanced analytics. It provides services like Strategy Consulting, Advisory Consulting and high quality classroom individual and corporate training.

See full profile