#Test - Create a Data Lakehouse in Kubernetes
-
Updated
May 25, 2024
#Test - Create a Data Lakehouse in Kubernetes
🦖 Efficiently evolve your old fixed-length data files into more modern file formats, fully parallelized!
This project implements an end-to-end techstack for a data platform, can be used on production.
Data lakehouse at home with docker compose
Qbeast-spark: DataSource enabling multi-dimensional indexing and efficient data sampling. Big Data, free from the unnecessary!
STEDI project
This repository is a place for the Data Warehousing course at the Information Systems & Analytics department, Santa Clara University.
The project aims to process Formula 1 racing data, create an automated data pipeline, and make the data available for presentation and analysis purposes.
My M.Sc. dissertation: Modern Data Platform using DataOps, Kubernetes, and Cloud-Native ecosystem to build a resilient Big Data platform based on Data Lakehouse architecture which is the base for Machine Learning (MLOps) and Artificial Intelligence (AIOps).
Sample Data Lakehouse deployed in Docker containers using Apache Iceberg, Minio, Trino and a Hive Metastore. Can be used for local testing.
A curated list of open source tools used in analytical stacks and data engineering ecosystem
Qbeast-spark: DataSource enabling multi-dimensional indexing and efficient data sampling. Big Data, free from the unnecessary!
Add a description, image, and links to the data-lakehouse topic page so that developers can more easily learn about it.
To associate your repository with the data-lakehouse topic, visit your repo's landing page and select "manage topics."