Learn Hadoop and Big Data

Rodney Rossi

Play Speed
  • 0.5x
  • 1x (Normal)
  • 1.25x
  • 1.5x
  • 2x
9 Lessons (9h 48m)
    • 1. 01 Introduction To Course

      3:06
    • 2. 2 Add Value to Existing Data with Mapreduce

      57:03
    • 3. 3 Hadoop Analytics and NoSQL

      54:38
    • 4. 4 Kafka Streaming with Yarn and Zookeeper

      59:57
    • 5. 5 Real Time Stream processing with Apache Kafka and Apache Storm

      58:46
    • 6. 6 Big Data Applications

      58:14
    • 7. 7 Log collection and analytics with the Hadoop Distributed

      57:37
    • 8. 8 Data Science with Hadoop Predictive Analytics

      44:15
    • 9. 9 Visual Analytics and Big Data Analytics for e commerce

      194:45

About This Class

Hadoop is an open source distributed processing framework that manages data processing and storage for big data applications in scalable clusters of computer servers. It's at the center of an ecosystem of big data technologies that are primarily used to support advanced analytics initiatives, including predictive analytics, data mining and machine learning.

Hadoop systems can handle various forms of structured and unstructured data, giving users more flexibility for collecting, processing and analyzing data than relational databases and data warehouses provide.

Prerequisite knowledge required:

  • Some activities will require some prior programming experience
  • A basic familiarity with the Linux command line will be very helpful.
  • You will need access to a PC running 64-bit Windows, MacOS, or Linux with an Internet connection

This class has these contents covered,
1. Introduction
2. Add Value to Existing Data with Mapreduce
3. Hadoop Analytics and NoSQL
4. Kafka Streaming with Yarn and Zookeeper
5. Real Time Stream processing with Apache Kafka and Apache Storm
6. Big Data Applications
7. Log collection and analytics with the Hadoop Distributed
8. Data Science with Hadoop Predictive Analytics
9. Visual Analytics and Big Data Analytics for e-commerce