Modeling Data Warehouses using Apache Hive

3.4
estrellas
5 calificaciones
ofrecido por
Coursera Project Network
En este proyecto guiado, tú:

Design a data warehouse

Build a data warehouse using Apache Hive over apache Hadoop

Query a data warehouse for specific records

Clock1 hour 40 minutes
BeginnerPrincipiante
CloudNo se necesita descarga
VideoVideo de pantalla dividida
Comment DotsInglés (English)
LaptopSolo escritorio

By the end of this project, you will create your own data warehouse by using a star schema modeling technique that holds baseball raw data gathered from multiple baseball leagues in the States. During the course of this project, you will be using Apache Hadoop and Apache Hive as the main tools to build your very own Data Warehouse using HQL as the querying language, which is very similar to SQL. The learner will be introduced to the star schema modeling technique, database creation and handling, data insertion, and data storage using two different methods in Apache Hadoop. Data Warehouses are viewed as the core of any application, as any further step that requires dealing with the data stored such as Data Analytics, building a recommendation system, or applying any sort of Machine Learning will inevitably point out to the start point, which is the data warehouse modeling. This guided project is for students who are aiming to start their careers in the broad field of Data Engineering and the specific field of Data Warehouses Modeling. It will also be of interest to people who are keen on exploring Big Data platforms such as Apache Hadoop and Apache Hive. Note: This project works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.

Habilidades que desarrollarás

Data Warehouse DevelopmentDatabase (DB) DesignData WarehousingDatabase Management Systems

Aprende paso a paso

En un video que se reproduce en una pantalla dividida con tu área de trabajo, tu instructor te guiará en cada paso:

  1. By the end of Task 1,you will be able to identify the star schema format in our project description.

  2. By the end of Task 2, you will be able to identify the attributes and their data types.

  3. By the end of Task 3, you will be able to create a database and tables in the database using Apache Hadoop and Apache Hive.

  4. By the end of Task 4, you will be able to insert data into tables manually.

  5. By the end of Task 5, you will be able to amend tables by adding primary and foreign keys constraints.

  6. By the end of Task 6,   you will be able to apply useful queries to your data warehouse to fetch specific data records.

Cómo funcionan los proyectos guiados

Tu espacio de trabajo es un escritorio virtual directamente en tu navegador, no requiere descarga.

En un video de pantalla dividida, tu instructor te guía paso a paso

Preguntas Frecuentes

Preguntas Frecuentes

¿Tienes más preguntas? Visita el Centro de Ayuda al Alumno.