New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Roadmap 2023 #26
Comments
Some feedback from the community
Team discussions
Feel free to add more. |
Support delete, update, user permission control stay the same the latest version of clickhouse database |
Support CREATE Table AS SELECT * syntax |
sql fingerprint support |
Other import support :RocketMQ, MaterializedMySQL, MaterializedPostgreSQL, Flink, Pulsar |
Make shared metadata of object storage to be compatible with JuiceFS, then we can shared checkpoint of cnch merge tree part with original ClickHouse using DiskLocal through mounted Juicefs. In some time-travel scan case, we can use latest original ClickHouse which has new feature that ByConity hav't. |
automatic collection, update and analysis of statistical information |
support distributed cache for higher query cache hit rate |
Support JuiceFS |
Byconity provides an all-in-one package. You can install clickhouse-client,clickhouse-server,clickhouse-worker,tso_server,daemon_manager,and resource_manager all at once with this unified package; You can install some components or specify a component version as required. |
Support RBAC,Support for SQL driven maintenance |
Clone table |
Can it support Apache Paimon? |
Roadmap updating proposal / 路线图更新提案 ===English version===
As mentioned above, the roadmap update consists of two parts: adding new features and removing or adjusting some old functionalities. Let's break down these two parts of the update.
With the addition of the aforementioned high-priority features, we have also removed or adjusted some old functionalities. These involve features with unclear requirements and code refactoring. The detailed list is as follows, and we will allocate time to support the removed functionalities.
Towards the end of each quarter, we conduct a review and fine-tune the content for the following quarter. We synchronize these adjustments with the community and welcome discussions, comments, and new feature requests. The finalized roadmap will be updated after the first week of each quarter and any newly proposed feature requests will be considered for the subsequent quarter. ===中文版===
如上所述,路线图的更新包括2部分,一部分是新增功能,一部分是移除和调整了部分旧的功能,这里对这2部分更新进行拆解。
由于新增了上述高优的功能,我们也移除和调整了部分旧的功能,这部分功能涉及到一些需求不明确的功能和代码重构,详细列表如下所示,对移除的功能会重新安排时间去支持。
我们每个季度临近结束的时候都会进行一次review和并对后续季度的内容进行微调,并把调整同步到社区,欢迎大家讨论评论和提新的功能需求,并在每个季度的第一周结束之后进行定稿,然后更新社区路线图。定稿之后提的新功能需求会顺延到下一个季度。 |
MetaData Backup and recover If the metadata kv FoundationDB is broken (for example, disk broken or logic fault cause FDB broken) and metadata is lost, is there any method to restore data?Therefore, I suggest adding the metadata backup and recovery function, which I hope to move to Q3 plan. |
The ByConity support ELT feature is an exciting feature and a valuable feature. |
Is inverted index also supported? In addition to the good performance of the primary key index in the current index, when querying non-primary key fields, the performance needs to be improved. |
CnchMergeTree support materialized view. |
I hope to support joint query engines for multiple tables and multiple data sources, pushing queries down to their respective data sources for querying. The joint query engine consolidates the data from each data source. Merge Table Engine similar to clickhouse |
Build Objectives
Requirements1. Data warehouse scenarioIn data warehouse scenarios, we focus on scenario coverage improvement and support complex query of large tables in data warehouses. Key Features1.1. Multi-stage execution and ETL capabilities at the execution layer, batch processing and exchange shufle are supported, and complex SQL statements for querying large tables in the data warehouse are supported. General Features1.4. Support multi external catalog; 2. OLAP Scenariowe focus on reliability and performance improvement in the OLAP scenario 3. Reliability enhancement3.1. Metadata backup and restoration capabilities What we foucs feature is 1.1, 1.2 and 3.1 目标
需求一、数仓场景数仓场景关注场景覆盖率提升和支持数仓的大表复杂查询 关键特性1.1. 执行层的多Stage执行、ETL能力,支持batch processing和exchange shufle; 通用特性1.4. Multi External catalog; 二、Olap多维分析场景Olap场景关注可靠性和性能提升 三、可靠性增强3.1. 元数据具备备份和恢复能力 |
Finished roadmap from Q1 and Q2 Storage
Index
Stability
Installation
CI
|
Support Apache Paimon,it‘s very cool Stream Data Lake!!! |
Updates on roadmap:
added:
|
Welcome to share your ideas on the roadmap. The updated roadmap for Q3 and Q4 are shown below.
Storage
feat: Enable data preload write through #189
feat: patch S3 storage basic #260
External Table/Data Lake (project https://github.com/orgs/ByConity/projects/2)
Hive table should support schema inference #315
Support reading Hudi table #360
Support external catalogs #361
Hive 2.0 #550
Hive orc/parquet min-max index pruning #551
RFC: Hive distributed processing #220
Index
perf: In-memory cache for primary key index #209
Runtime
Asynchronous query execution、query Queue、join spill
Exchange spill、colocated scheduling、batch execution
ByConity Support UDF #427
Optimizer
Transaction
Enterprise feature
Performance improvement
Stability
RFC: Multi-Server Integration #208
chore: Improve CAS operation #145
use atomic api for compare and clear #185
Installation
add script to build debian and rpm package #100
https://github.com/ByConity/ByConity/tree/master/packages
CI
add testing guideline #206
The text was updated successfully, but these errors were encountered: