Bruin is a data pipeline tool that is designed to be easy-to-use. It allows building data pipelines using SQL and Python, and has built-in data quality checks.
-
Updated
May 14, 2024 - Go
Bruin is a data pipeline tool that is designed to be easy-to-use. It allows building data pipelines using SQL and Python, and has built-in data quality checks.
🧙 Build, run, and manage data pipelines for integrating and transforming data.
Apache DolphinScheduler is the modern data orchestration platform. Agile to create high performance workflow with low-code
Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Database replication platform that leverages change data capture. Stream production data from databases to your data warehouse (Snowflake, BigQuery, Redshift) in real-time.
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
An orchestration platform for the development, production, and observation of data assets.
The framework for fast development and deployment of RAG systems.
Lean and mean distributed stream processing system written in rust and web assembly.
Meltano: the declarative code-first data integration engine that powers your wildest data and ML-powered product ideas. Say goodbye to writing, maintaining, and scaling your own API integrations.
Framework for standardizing, transforming, and applying quality checks to time series data.
Building data processing pipelines for documents processing with NLP using Apache NiFi and related services
One framework to develop, deploy and operate data workflows with Python and SQL.
dbt package that is part of Elementary, the dbt-native data observability solution for data & analytics engineers. Monitor your data pipelines in minutes. Available as self-hosted or cloud service with premium features.
The dbt-native data observability solution for data & analytics engineers. Monitor your data pipelines in minutes. Available as self-hosted or cloud service with premium features.
Move your data with ease.
Every thing about designing installing and implementing data pipelines to include kafka zookeeper hadoop If you enjoy my content please consider supporting what I do Thank you.
A Realtime Seismic Logging & Alerts Service with Live Monitoring & Email Alerts made using Kafka Data Pipelines, all Dockerized & Deployment Ready!
Main repo including core data model, data marts, reference data, terminology, and the clinical concept library
Add a description, image, and links to the data-pipelines topic page so that developers can more easily learn about it.
To associate your repository with the data-pipelines topic, visit your repo's landing page and select "manage topics."