SeaTunnel is a distributed, high-performance data integration platform for the synchronization and transformation of massive data (offline & real-time).
-
Updated
Jun 13, 2024 - Java
SeaTunnel is a distributed, high-performance data integration platform for the synchronization and transformation of massive data (offline & real-time).
Apply Data Engineering to Personal Finance
The leading data integration platform for ETL / ELT data pipelines from APIs, databases & files to data warehouses, data lakes & data lakehouses. Both self-hosted and Cloud-hosted.
Feldera Continuous Analytics Platform
Toolkit for describing data transformation pipelines by compositing simple reusable components.
Conduit streams data between data stores. Kafka Connect replacement. No JVM required.
Infinitely scalable, event-driven, language-agnostic orchestration and scheduling platform to manage millions of workflows declaratively in code.
ingestr is a CLI tool to copy data between any databases with a single command seamlessly.
Sample data set exemplifying an idealized data processing pipeline for didactic purposes
Privacy and Security focused Segment-alternative, in Golang and React
使用ETL data pipeline 將UBER 資料清洗、排程、最後放置在GCP上運行與後續分析 的專案
Watchmen Platform is a low code data platform for data pipeline, meta data management , analysis, indicator objective analysis and quality management
Cryptocurrency prediction using LSTM (Long Short Term Memory)
Aqueduct Core is responsible for the core functionality of Aqueduct, an experiment management system.
Flink CDC is a streaming data integration tool
Pythonic tool for orchestrating machine-learning/high performance/quantum-computing workflows in heterogeneous compute environments.
Solução completa dedicada a realizar ETL de dados de cotações de moedas usando Python. Fonte dos dados: https://docs.awesomeapi.com.br/api-de-moedas
Jayvee is a domain-specific language and runtime for automated processing of data pipelines
The dbt-native data observability solution for data & analytics engineers. Monitor your data pipelines in minutes. Available as self-hosted or cloud service with premium features.
Add a description, image, and links to the data-pipeline topic page so that developers can more easily learn about it.
To associate your repository with the data-pipeline topic, visit your repo's landing page and select "manage topics."