-
Notifications
You must be signed in to change notification settings - Fork 286
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
data lost after scale-in 3x tikv node #2400
Comments
Could you upload TiCDC and TiKV logs? |
tikv and cdc logs have been uploaded to qa-minio/nfs/CDC_issue_log/CDC_issue_log/2400 |
Update: The root cause is still unknown, here is what I found for now.
The above shows one of the lost row
So the data lose should occurs between |
Current status
|
We have fixed several correctness bugs e.g., #3545 , and it does not happen after since, change to severity/moderate |
We also have fixed in TiKV side that may cause data lose, e.g., #4782. I'm going to close the issue, feel free to reopen if it happens again. |
Bug Report
Please answer these questions before submitting your issue. Thanks!
So TiKV node quantity totally is 12, and 2x CDC server, 2x changefeed, workload write to upstream tidb cluster:
$./bin/go-ycsb load mysql -P ./workloads/oncall2929 -p mysql.host=172.16.6.25 -p mysql.port=4000 -p operationcount=5000000 -p autocommit=1 --threads=200 -p databasename=test123 -p tablename=test0
Starting component
cdc
: /root/.tiup/components/cdc/v4.0.14/cdc cli changefeed list --pd=http://172.16.6.25:2379[
{
"id": "kafka-task-23",
"summary": {
"state": "normal",
"tso": 426624137162391554,
"checkpoint": "2021-07-28 11:12:09.483",
"error": null
}
},
{
"id": "replication-task-24",
"summary": {
"state": "normal",
"tso": 426624137162391554,
"checkpoint": "2021-07-28 11:12:09.483",
"error": null
}
}
]
scale-in 3x tikv node: tiup cluster scale-in 360UP -N 172.16.6.12:20170, tiup cluster scale-in 360UP -N 172.16.6.13:20171, tiup cluster scale-in 360UP -N 172.16.6.13:20170
Check tiup audit log:
fGcys7dWFBX:%2Froot%2F.tiup%2Fcomponents%2Fcluster%2Fv1.5.1%2Ftiup-cluster scale-in 360UP -N 172.16.6.12%3A20170
fGcys7dWFBX:2021-07-28T11:15:03.550+0800 INFO Execute command {"command": "tiup cluster scale-in 360UP -N 172.16.6.12:20170"}
fGcyt3HVY3p:%2Froot%2F.tiup%2Fcomponents%2Fcluster%2Fv1.5.1%2Ftiup-cluster scale-in 360UP -N 172.16.6.13%3A20171
fGcyt3HVY3p:2021-07-28T11:15:19.819+0800 INFO Execute command {"command": "tiup cluster scale-in 360UP -N 172.16.6.13:20171"}
fGcyvgT9nRj:%2Froot%2F.tiup%2Fcomponents%2Fcluster%2Fv1.5.1%2Ftiup-cluster scale-in 360UP -N 172.16.6.13%3A20170
fGcyvgT9nRj:2021-07-28T11:15:38.292+0800 INFO Execute command {"command": "tiup cluster scale-in 360UP -N 172.16.6.13:20170"}
Stop workload I/O, waiting for sync task completing.
Check data consistency.
upstream 6.24:
MySQL [test123]> select count() from test0;
+-----------+
| count() |
+-----------+
| 333895176 |
+-----------+
1 row in set (8.26 sec)
downstream 6.30:
MySQL [test123]> select count() from test0;
+-----------+
| count() |
+-----------+
| 333895126 |
+-----------+
1 row in set (8.96 sec)
$curl http://172.16.6.30:10080/mvcc/key/test123/test0\(p20210508\)/16585000 ->"start_ts": 426624261094637572
$ tiup ctl:v4.0.14 pd -u http://72.16.6.30:2379 tso 426624261094637572 ->system: 2021-07-28 11:20:02.247 +0800 CST
Versions of the cluster
Upstream TiDB cluster version (execute
SELECT tidb_version();
in a MySQL client):TiCDC version (execute
cdc version
):The text was updated successfully, but these errors were encountered: