Modeling Data Warehouses using Apache Hive

4.2
звезд
Оценки: 15
от партнера
Coursera Project Network
В этом Проект с консультациями вы:

Design a data warehouse

Build a data warehouse using Apache Hive over apache Hadoop

Query a data warehouse for specific records

Clock1 hour 40 minutes
BeginnerНачинающий
CloudЗагрузка не требуется
VideoВидео на разделенном экране
Comment DotsАнглийский
LaptopТолько для ПК

By the end of this project, you will create your own data warehouse by using a star schema modeling technique that holds baseball raw data gathered from multiple baseball leagues in the States. During the course of this project, you will be using Apache Hadoop and Apache Hive as the main tools to build your very own Data Warehouse using HQL as the querying language, which is very similar to SQL. The learner will be introduced to the star schema modeling technique, database creation and handling, data insertion, and data storage using two different methods in Apache Hadoop. Data Warehouses are viewed as the core of any application, as any further step that requires dealing with the data stored such as Data Analytics, building a recommendation system, or applying any sort of Machine Learning will inevitably point out to the start point, which is the data warehouse modeling. This guided project is for students who are aiming to start their careers in the broad field of Data Engineering and the specific field of Data Warehouses Modeling. It will also be of interest to people who are keen on exploring Big Data platforms such as Apache Hadoop and Apache Hive. Note: This project works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.

Навыки, которые вы получите

Data Warehouse DevelopmentDatabase (DB) DesignData WarehousingDatabase Management Systems

Будете учиться пошагово

На видео, которое откроется рядом с рабочей областью, преподаватель объяснит эти шаги:

  1. By the end of Task 1,you will be able to identify the star schema format in our project description.

  2. By the end of Task 2, you will be able to identify the attributes and their data types.

  3. By the end of Task 3, you will be able to create a database and tables in the database using Apache Hadoop and Apache Hive.

  4. By the end of Task 4, you will be able to insert data into tables manually.

  5. By the end of Task 5, you will be able to amend tables by adding primary and foreign keys constraints.

  6. By the end of Task 6,   you will be able to apply useful queries to your data warehouse to fetch specific data records.

Как устроены проекты с консультациями

Ваше рабочее пространство — это облачный рабочий стол в браузере. Ничего не нужно загружать.

На разделенном экране видео преподаватель предоставляет пошаговые

Рецензии

Лучшие отзывы о курсе MODELING DATA WAREHOUSES USING APACHE HIVE

Посмотреть все отзывы

Часто задаваемые вопросы

Часто задаваемые вопросы

Остались вопросы? Посетите Центр поддержки учащихся.