Modeling Data Warehouses using Apache Hive

4.1
14 個評分
提供方
Coursera Project Network
在此指導 項目中,您將:

Design a data warehouse

Build a data warehouse using Apache Hive over apache Hadoop

Query a data warehouse for specific records

Clock1 hour 40 minutes
Beginner初級
Cloud無需下載
Video分屏視頻
Comment Dots英語(English)
Laptop僅限桌面

By the end of this project, you will create your own data warehouse by using a star schema modeling technique that holds baseball raw data gathered from multiple baseball leagues in the States. During the course of this project, you will be using Apache Hadoop and Apache Hive as the main tools to build your very own Data Warehouse using HQL as the querying language, which is very similar to SQL. The learner will be introduced to the star schema modeling technique, database creation and handling, data insertion, and data storage using two different methods in Apache Hadoop. Data Warehouses are viewed as the core of any application, as any further step that requires dealing with the data stored such as Data Analytics, building a recommendation system, or applying any sort of Machine Learning will inevitably point out to the start point, which is the data warehouse modeling. This guided project is for students who are aiming to start their careers in the broad field of Data Engineering and the specific field of Data Warehouses Modeling. It will also be of interest to people who are keen on exploring Big Data platforms such as Apache Hadoop and Apache Hive. Note: This project works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.

您要培養的技能

Data Warehouse DevelopmentDatabase (DB) DesignData WarehousingDatabase Management Systems

分步進行學習

在與您的工作區一起在分屏中播放的視頻中,您的授課教師將指導您完成每個步驟:

  1. By the end of Task 1,you will be able to identify the star schema format in our project description.

  2. By the end of Task 2, you will be able to identify the attributes and their data types.

  3. By the end of Task 3, you will be able to create a database and tables in the database using Apache Hadoop and Apache Hive.

  4. By the end of Task 4, you will be able to insert data into tables manually.

  5. By the end of Task 5, you will be able to amend tables by adding primary and foreign keys constraints.

  6. By the end of Task 6,   you will be able to apply useful queries to your data warehouse to fetch specific data records.

指導項目工作原理

您的工作空間就是瀏覽器中的雲桌面,無需下載

在分屏視頻中,您的授課教師會為您提供分步指導

審閱

來自MODELING DATA WAREHOUSES USING APACHE HIVE的熱門評論

查看所有評論

常見問題

常見問題

還有其他問題嗎?請訪問 學生幫助中心