-
Notifications
You must be signed in to change notification settings - Fork 226
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEATURE] properly backup S3/GCS/Azure disks, support server-side copy to back up bucket during clickhouse-backup create
and during clickhouse-backup restore
#447
Comments
Unfortunately, s3 disk not full supported by During adding s3 support we didn't properly clean data in test case, and currently OK in test is false positive Now clickhouse-backup successful backup only metadata which contains S3 keys, not real data So, currently, I don't know the good solution to full support S3 disks.
|
@Slach thanks for your explanation. should I close the ticket or let it remain open for future improvement? |
want to confirm one more thing: if everything go smoothly, partition name should remain intact after restoring from backup, right? I ask above question because I saw while restoring, partition name from default disk changed |
what exatcly you mean when say "partition name" ? |
By "partition name", I mean 'name' column from above SQL. In source clickhouse, the partition name in default disk changes from '2022_6_6_0' to '2022_1_1_0' in restored clickhouse |
this is not "data partition", this is "data part" I don't know, shall part name changed after
merge_level = 0 means this part never merged with another parts, so block number 6 was never merged with 5 and 4 |
clickhouse-backup create
and properly restore storage policy
@Slach do we have an estimated time for this feature to be ready? thanks |
no estimate with this feature now, on v2 branch we already implements you can try to use rclone sync with v2 branch for feel free to make pull request |
thanks for explanation @Slach |
Perhaps it would be nice to add a switch for this feature (snapshot data on S3 disk)? For my use case, I need to back up metadata and metadata only. |
@gfunc if you want to back up metadata (it means database \table \ views \ funcitons SQL definitions only) just use |
clickhouse-backup create
and properly restore storage policyclickhouse-backup create
and properly restore with check storage policy
clickhouse-backup create
and properly restore with check storage policyclickhouse-backup create
and properly restore with check storage policy
clickhouse-backup create
and properly restore with check storage policyclickhouse-backup create
and during clickhouse-backup restore
@Slach I need to back up metadata only, but when I used option |
@silentsokolov could you create separate issue |
I have two disks: one is default, another is s3, here is the storage configuration both in source clickhouse and target clickhouse:
I created a table using storage policy 'tiered' so that I can store part of the data into s3 bucket.
When I restore the clickhouse db using clickhouse-backup, db partitions on default disk are properly restored, but db partitions on s3 disk cannot be restored. Throwing exceptions while running commands:
ALTER TABLE fedex.ontime ATTACH PART '2022_2_2_0';
Code: 499. DB::Exception: Received from localhost:9000. DB::Exception: The specified key does not exist.. (S3_ERROR)
In target clickhouse db, I have the following detached parts:
please help, thanks in advance
The text was updated successfully, but these errors were encountered: