Skip to content

BrightOsas/NYC-Taxi-Trip-Data-Pipeline

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

55 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DATA ENGINEERING PROJECT

TITLE: Building an Efficient Batch Data Pipeline for Analytical Insights

Introduction

Welcome to my first Data Engineering project! The objective is to construct an efficient data pipeline for improved data management and more effective data-driven suggestions. This project involves both ETL and ELT operations, showcasing my understanding of how to utilize big data tools to build an efficient data pipeline.

r2

Architecture

taxi-architecture drawio

ELT/ETL Operations Overview

ELT Operations: Data is extracted from the NYC trip website, loaded into a PostgreSQL database, and transformed using DBT.

ETL Operations: The transformed DBT data is extracted from the PostgreSQL database using PySpark, undergoes further transformation, and is then loaded into another PostgreSQL database for virtualization. This process includes aggregating the data to reduce the number of rows from over 18 million to 2000-plus.

Data Storage: Transformed DBT data is stored in an AWS S3 bucket using PySpark and Boto3.

Dashboard

Data Description: Dataset contains data for the year 2023, it provides insight on the Taxi trips record for the 6 boroughs in New York City with respect to taxi vendors, service type and trip record type.

Link To Dashboard

NEW

db3

Technologies

  • Docker: Containerization
  • Apache Airflow: Orchestration
  • Terraform: Infrastructure as Code (IaaC)
  • PostgreSQL: Database and Data Warehousing
  • DBT (Data Build Tool) Local: Data Transformation and Batch Processing
  • PySpark: Batch Processing and Data Transformation
  • Looker Studio: Dashboard Virtualization
  • Amazon S3: Data Lake
  • Boto3: Batch Processing
  • Python: Scripting

How to run the Project

clone this repository
set up the environment and requiremnt

  • cd data-ingestion folder : docker compose up -d --build
  • cd airflow orchestration folder : docker compose up -d --build
  • setup dbt(local) python env: dbt run
  • run pyspark script in jupyter environment.

contact

About

DataTalkClub Data engineering Bootcamp Project: Building an Efficient Batch Data Pipeline for Analytical Insights

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published