Working with Big Data

提供方
Coursera Project Network
在此指導項目中,您將:

Process a large dataset from NOAA showing hourly precipitation rates for a ten year period from the state of Wisconsin

Clock2 hours
Intermediate中級
Cloud無需下載
Video分屏視頻
Comment Dots英語(English)
Laptop僅限桌面

By the end of this project, you will set up an environment for Big Data Development using Visual Studio Code, MongoDB and Apache Spark. You will then use the environment to process a large dataset from NOAA showing hourly precipitation rates for a ten year period from the state of Wisconsin. MongoDB is a widely used NoSQL database well suited for very large datasets or Big Data. It is highly scalable and adaptable as well. Apache Spark is used for efficient in-memory processing of Big Data.

您要培養的技能

  • PySpark Queries
  • Mongodb
  • Python Programming
  • Big Data
  • PySpark

分步進行學習

在與您的工作區一起在分屏中播放的視頻中,您的授課教師將指導您完成每個步驟:

  1. Set up Apache Spark and MongoDB Environment.

  2. Create a Python PySpark program to read CSV data.

  3. Use Spark SQL to query in-memory data.

  4. Configure Apache Spark to connect to MongoDB.

  5. Persist data using Spark and MongoDB.

指導項目工作原理

您的工作空間就是瀏覽器中的雲桌面,無需下載

在分屏視頻中,您的授課教師會為您提供分步指導

常見問題

常見問題

還有其他問題嗎?請訪問 學生幫助中心