Skip to content

Phylake1337/uber-data-pipeline-end2end

Repository files navigation

Uber Data Analytics | Data Engineering GCP Project

Introduction

The goal of this project is to perform data analytics on Uber data using various tools and technologies, including GCP Storage, Python, Compute Instance, Mage Data Pipeline Tool, BigQuery, and Looker Studio.

Implementation Details

  1. Layout data pipeline architecture.

  1. Host Uber data (CSV file) on staging storage (google storage).

  2. Set up Google Compute Instance (VM) with Python and Mage to handle the ETL process.

  3. Model the data into various tables (star schema).

  1. Write Python scripts on Mage to:
  • Extract data from google cloud.
  • Transform, filter, and split the data into multiple tables.
  • Load the data into BigQuery schema.

  1. Create a new analytics table to feed Looker Dashbaord.

  2. Set up Looker Dashbaord to visualize the data into different charts.

Technology Used

  • Programming Language - Python

Google Cloud Platform

  1. Google Storage
  2. Compute Instance
  3. BigQuery
  4. Looker Studio

Dataset Used

TLC Trip Record Data Yellow and green taxi trip records include fields capturing pick-up and drop-off dates/times, pick-up and drop-off locations, trip distances, itemized fares, rate types, payment types, and driver-reported passenger counts.

More info about the dataset can be found here:

  1. Website
  2. Data Dictionary

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published