Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

syncer(dm): use an early location to reset binlog and open safemode #3860

Merged
merged 5 commits into from
Dec 22, 2021

Conversation

lance6716
Copy link
Contributor

@lance6716 lance6716 commented Dec 13, 2021

What problem does this PR solve?

close https://github.com/pingcap/ticdc/issues/3487 for release-5.3
https://github.com/pingcap/ticdc/issues/3711 of master branch need a better solution

What is changed and how it works?

we use globalCheckpoint which is a correct-but-early location to reset binlog streamer, and open one transaction safemode

add a half of an integration test back

Check List

Tests

  • Integration test

Code changes

Side effects

Related changes

Release note

Fix a bug of data loss when DM does finer grained retry

@ti-chi-bot
Copy link
Member

ti-chi-bot commented Dec 13, 2021

[REVIEW NOTIFICATION]

This pull request has been approved by:

  • glorv
  • lichunzhu

To complete the pull request process, please ask the reviewers in the list to review by filling /cc @reviewer in the comment.
After your PR has acquired the required number of LGTMs, you can assign this pull request to the committer in the list by filling /assign @committer in the comment to help you merge this pull request.

The full list of commands accepted by this bot can be found here.

Reviewer can indicate their review by submitting an approval review.
Reviewer can cancel approval by submitting a request changes review.

@ti-chi-bot ti-chi-bot added release-note Denotes a PR that will be considered when it comes time to generate release notes. do-not-merge/cherry-pick-not-approved size/M Denotes a PR that changes 30-99 lines, ignoring generated files. labels Dec 13, 2021
@lance6716 lance6716 added the area/dm Issues or PRs related to DM. label Dec 13, 2021
@codecov-commenter
Copy link

Codecov Report

❗ No coverage uploaded for pull request base (release-5.3@961af87). Click here to learn what that means.
The diff coverage is n/a.

@@               Coverage Diff                @@
##             release-5.3      #3860   +/-   ##
================================================
  Coverage               ?   56.6504%           
================================================
  Files                  ?        212           
  Lines                  ?      22946           
  Branches               ?          0           
================================================
  Hits                   ?      12999           
  Misses                 ?       8627           
  Partials               ?       1320           

@lance6716
Copy link
Contributor Author

/run-all-tests

if err != nil {
return err
}
log.L().Info("reset replication binlog puller", zap.Any("pos", s.checkpoint.GlobalPoint()))
_ = s.safeMode.Add(tctx, 1)
Copy link
Contributor

@glorv glorv Dec 15, 2021

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How about use a local variable like inFinerRetry to record this temporary safe-mode

safeMode: s.safeMode.Enable() || retrySafeMode,

@@ -1621,11 +1632,13 @@ func (s *Syncer) Run(ctx context.Context) (err error) {

if s.streamerController.CanRetry(err) {
// lastLocation is the last finished GTID
err = s.streamerController.ResetReplicationSyncer(tctx, lastLocation)
err = s.streamerController.ResetReplicationSyncer(tctx, s.checkpoint.GlobalPoint())
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If this retry is occurred at sharding-resyc phase, enable safe-mode for one transaction is not enough because the start point may be reset to the very beginning point. But since safe-mode is enable in the full sharding-resync phase this is ok. Maybe we can add a comment about this here.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What does safe-mode is enable in the full sharding-resync phase means? To me this is a problem.

Copy link
Contributor Author

@lance6716 lance6716 Dec 15, 2021

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good to me

Copy link
Contributor

@lichunzhu lichunzhu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Rest LGTM

@@ -1621,11 +1632,13 @@ func (s *Syncer) Run(ctx context.Context) (err error) {

if s.streamerController.CanRetry(err) {
// lastLocation is the last finished GTID
err = s.streamerController.ResetReplicationSyncer(tctx, lastLocation)
err = s.streamerController.ResetReplicationSyncer(tctx, s.checkpoint.GlobalPoint())
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What does safe-mode is enable in the full sharding-resync phase means? To me this is a problem.

@lance6716
Copy link
Contributor Author

ping @lichunzhu @glorv

Copy link
Contributor

@glorv glorv left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@ti-chi-bot ti-chi-bot added the status/LGT1 Indicates that a PR has LGTM 1. label Dec 21, 2021
@lance6716
Copy link
Contributor Author

/cc @GMHDBJD

@ti-chi-bot ti-chi-bot added status/LGT2 Indicates that a PR has LGTM 2. and removed status/LGT1 Indicates that a PR has LGTM 1. labels Dec 21, 2021
@lance6716
Copy link
Contributor Author

/assign @nongfushanquan

@lance6716
Copy link
Contributor Author

/merge

@ti-chi-bot
Copy link
Member

This pull request has been accepted and is ready to merge.

Commit hash: 28ef93a

@ti-chi-bot ti-chi-bot added the status/can-merge Indicates a PR has been approved by a committer. label Dec 22, 2021
@ti-chi-bot
Copy link
Member

@lance6716: Your PR was out of date, I have automatically updated it for you.

At the same time I will also trigger all tests for you:

/run-all-tests

If the CI test fails, you just re-trigger the test that failed and the bot will merge the PR for you after the CI passes.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the ti-community-infra/tichi repository.

@lance6716
Copy link
Contributor Author

/run-leak-test

@ti-chi-bot ti-chi-bot merged commit 167963a into pingcap:release-5.3 Dec 22, 2021
overvenus pushed a commit that referenced this pull request Jan 18, 2022
* fix the txn_batch_size metric inaccuracy bug when the sink target is MQ

* address comments

* add comments for exported functions

* fix the compiling problem

* workerpool: limit the rate to output deadlock warning (#3775) (#3795)

* tests(ticdc): set up the sync diff output directory correctly (#3725) (#3741)

* relay(dm): use binlog name comparison (#3710) (#3712)

* dm/load: fix concurrent call Loader.Status (#3459) (#3468)

* cdc/sorter: make unified sorter cgroup aware (#3436) (#3439)

* tz (ticdc): fix timezone error (#3887) (#3906)

* pkg,cdc: do not use log package (#3902) (#3940)

* *: rename repo from pingcap/ticdc to pingcap/tiflow (#3959)

* http_*: add log for http api and refine the err handle logic (#2997) (#3307)

* etcd_worker: batch etcd patch (#3277) (#3389)

* http_api (ticdc): check --cert-allowed-cn before add server common name (#3628) (#3882)

* kvclient(ticdc): fix kvclient takes too long time to recover (#3612) (#3663)

* owner: fix owner tick block http request (#3490) (#3530)

* dm/syncer: use downstream PK/UK to generate DML (#3168) (#3256)

* dep(dm): update go-mysql (#3914) (#3934)

* dm/syncer: multiple rows use downstream schema (#3308) (#3953)

* errorutil,sink,syncer: add errorutil to handle ignorable error (#3264) (#3995)

* dm/worker: don't exit when failed to read checkpoint in relay (#3345) (#4005)

* syncer(dm): use an early location to reset binlog and open safemode (#3860)

* ticdc/owner: Fix ddl special comment syntax error (#3845) (#3978)

* dm/scheduler: fix inconsistent of relay status (#3474) (#4009)

* owner,scheduler(cdc): fix nil pointer panic in owner scheduler (#2980) (#4007) (#4016)

* config(ticdc): Fix old value configuration check for maxwell protocol (#3747) (#3783)

* sink(ticdc): cherry pick sink bug fix to release 5.3 (#4083)

* master(dm): clean and treat invalid load task (#4004) (#4145)

* loader: fix wrong progress in query-status for loader (#4093) (#4143)

close #3252

* ticdc/processor: Fix backoff base delay misconfiguration (#3992) (#4028)

* dm: load table structure from dump files (#3295) (#4163)

* compactor: fix duplicate entry in safemode (#3432) (#3434) (#4088)

* kv(ticdc): reduce eventfeed rate limited log (#4072) (#4111)

close #4006

* metrics(ticdc): add resolved ts and add changefeed to dataflow (#4038) (#4104)

* This is an automated cherry-pick of #4192

Signed-off-by: ti-chi-bot <ti-community-prow-bot@tidb.io>

* retry(dm): align with tidb latest error message (#4172) (#4254)

close #4159, close #4246

* owner(ticdc): Add bootstrap and try to fix the meta information in it (#3838) (#3865)

* redolog: add a precleanup process when s3 enable (#3525) (#3878)

* ddl(dm): make skipped ddl pass `SplitDDL()` (#4176) (#4227)

close #4173

* cdc/sink: remove Initialize method from the sink interface (#3682) (#3765)

Co-authored-by: Ling Jin <7138436+3AceShowHand@users.noreply.github.com>

* http_api (ticdc): fix http api 'get processor' panic. (#4117) (#4123)

close #3840

* sink (ticdc): fix a deadlock due to checkpointTs fall back in sinkNode (#4084) (#4099)

close #4055

* cdc/sink: adjust kafka initialization logic (#3192) (#4162)

* try fix conflicts.

* This is an automated cherry-pick of #4192

Signed-off-by: ti-chi-bot <ti-community-prow-bot@tidb.io>

* fix conflicts.

* fix conflicts.

Co-authored-by: zhaoxinyu <zhaoxinyu512@gmail.com>
Co-authored-by: amyangfei <yangfei@pingcap.com>
Co-authored-by: lance6716 <lance6716@gmail.com>
Co-authored-by: sdojjy <sdojjy@qq.com>
Co-authored-by: Ling Jin <7138436+3AceShowHand@users.noreply.github.com>
Co-authored-by: 3AceShowHand <jinl1037@hotmail.com>
@purelind purelind added this to the v5.3.1 milestone Feb 24, 2022
@lance6716 lance6716 deleted the manually-fix branch October 13, 2022 08:38
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/dm Issues or PRs related to DM. cherry-pick-approved Cherry pick PR approved by release team. release-note Denotes a PR that will be considered when it comes time to generate release notes. size/M Denotes a PR that changes 30-99 lines, ignoring generated files. status/can-merge Indicates a PR has been approved by a committer. status/LGT2 Indicates that a PR has LGTM 2.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

7 participants