-
Notifications
You must be signed in to change notification settings - Fork 287
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
syncer(dm): use an early location to reset binlog and open safemode #3860
Conversation
[REVIEW NOTIFICATION] This pull request has been approved by:
To complete the pull request process, please ask the reviewers in the list to review by filling The full list of commands accepted by this bot can be found here. Reviewer can indicate their review by submitting an approval review. |
Codecov Report
@@ Coverage Diff @@
## release-5.3 #3860 +/- ##
================================================
Coverage ? 56.6504%
================================================
Files ? 212
Lines ? 22946
Branches ? 0
================================================
Hits ? 12999
Misses ? 8627
Partials ? 1320 |
/run-all-tests |
dm/syncer/syncer.go
Outdated
if err != nil { | ||
return err | ||
} | ||
log.L().Info("reset replication binlog puller", zap.Any("pos", s.checkpoint.GlobalPoint())) | ||
_ = s.safeMode.Add(tctx, 1) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
How about use a local variable like inFinerRetry
to record this temporary safe-mode
safeMode: s.safeMode.Enable() || retrySafeMode,
@@ -1621,11 +1632,13 @@ func (s *Syncer) Run(ctx context.Context) (err error) { | |||
|
|||
if s.streamerController.CanRetry(err) { | |||
// lastLocation is the last finished GTID | |||
err = s.streamerController.ResetReplicationSyncer(tctx, lastLocation) | |||
err = s.streamerController.ResetReplicationSyncer(tctx, s.checkpoint.GlobalPoint()) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If this retry is occurred at sharding-resyc
phase, enable safe-mode for one transaction is not enough because the start point may be reset to the very beginning point. But since safe-mode is enable in the full sharding-resync phase this is ok. Maybe we can add a comment about this here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What does safe-mode is enable in the full sharding-resync phase
means? To me this is a problem.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we will open safemode when there's an unsynced sharding table
please help check it
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good to me
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Rest LGTM
@@ -1621,11 +1632,13 @@ func (s *Syncer) Run(ctx context.Context) (err error) { | |||
|
|||
if s.streamerController.CanRetry(err) { | |||
// lastLocation is the last finished GTID | |||
err = s.streamerController.ResetReplicationSyncer(tctx, lastLocation) | |||
err = s.streamerController.ResetReplicationSyncer(tctx, s.checkpoint.GlobalPoint()) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What does safe-mode is enable in the full sharding-resync phase
means? To me this is a problem.
ping @lichunzhu @glorv |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
/cc @GMHDBJD |
/assign @nongfushanquan |
/merge |
This pull request has been accepted and is ready to merge. Commit hash: 28ef93a
|
@lance6716: Your PR was out of date, I have automatically updated it for you. At the same time I will also trigger all tests for you: /run-all-tests If the CI test fails, you just re-trigger the test that failed and the bot will merge the PR for you after the CI passes. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the ti-community-infra/tichi repository. |
/run-leak-test |
* fix the txn_batch_size metric inaccuracy bug when the sink target is MQ * address comments * add comments for exported functions * fix the compiling problem * workerpool: limit the rate to output deadlock warning (#3775) (#3795) * tests(ticdc): set up the sync diff output directory correctly (#3725) (#3741) * relay(dm): use binlog name comparison (#3710) (#3712) * dm/load: fix concurrent call Loader.Status (#3459) (#3468) * cdc/sorter: make unified sorter cgroup aware (#3436) (#3439) * tz (ticdc): fix timezone error (#3887) (#3906) * pkg,cdc: do not use log package (#3902) (#3940) * *: rename repo from pingcap/ticdc to pingcap/tiflow (#3959) * http_*: add log for http api and refine the err handle logic (#2997) (#3307) * etcd_worker: batch etcd patch (#3277) (#3389) * http_api (ticdc): check --cert-allowed-cn before add server common name (#3628) (#3882) * kvclient(ticdc): fix kvclient takes too long time to recover (#3612) (#3663) * owner: fix owner tick block http request (#3490) (#3530) * dm/syncer: use downstream PK/UK to generate DML (#3168) (#3256) * dep(dm): update go-mysql (#3914) (#3934) * dm/syncer: multiple rows use downstream schema (#3308) (#3953) * errorutil,sink,syncer: add errorutil to handle ignorable error (#3264) (#3995) * dm/worker: don't exit when failed to read checkpoint in relay (#3345) (#4005) * syncer(dm): use an early location to reset binlog and open safemode (#3860) * ticdc/owner: Fix ddl special comment syntax error (#3845) (#3978) * dm/scheduler: fix inconsistent of relay status (#3474) (#4009) * owner,scheduler(cdc): fix nil pointer panic in owner scheduler (#2980) (#4007) (#4016) * config(ticdc): Fix old value configuration check for maxwell protocol (#3747) (#3783) * sink(ticdc): cherry pick sink bug fix to release 5.3 (#4083) * master(dm): clean and treat invalid load task (#4004) (#4145) * loader: fix wrong progress in query-status for loader (#4093) (#4143) close #3252 * ticdc/processor: Fix backoff base delay misconfiguration (#3992) (#4028) * dm: load table structure from dump files (#3295) (#4163) * compactor: fix duplicate entry in safemode (#3432) (#3434) (#4088) * kv(ticdc): reduce eventfeed rate limited log (#4072) (#4111) close #4006 * metrics(ticdc): add resolved ts and add changefeed to dataflow (#4038) (#4104) * This is an automated cherry-pick of #4192 Signed-off-by: ti-chi-bot <[email protected]> * retry(dm): align with tidb latest error message (#4172) (#4254) close #4159, close #4246 * owner(ticdc): Add bootstrap and try to fix the meta information in it (#3838) (#3865) * redolog: add a precleanup process when s3 enable (#3525) (#3878) * ddl(dm): make skipped ddl pass `SplitDDL()` (#4176) (#4227) close #4173 * cdc/sink: remove Initialize method from the sink interface (#3682) (#3765) Co-authored-by: Ling Jin <[email protected]> * http_api (ticdc): fix http api 'get processor' panic. (#4117) (#4123) close #3840 * sink (ticdc): fix a deadlock due to checkpointTs fall back in sinkNode (#4084) (#4099) close #4055 * cdc/sink: adjust kafka initialization logic (#3192) (#4162) * try fix conflicts. * This is an automated cherry-pick of #4192 Signed-off-by: ti-chi-bot <[email protected]> * fix conflicts. * fix conflicts. Co-authored-by: zhaoxinyu <[email protected]> Co-authored-by: amyangfei <[email protected]> Co-authored-by: lance6716 <[email protected]> Co-authored-by: sdojjy <[email protected]> Co-authored-by: Ling Jin <[email protected]> Co-authored-by: 3AceShowHand <[email protected]>
What problem does this PR solve?
close https://github.com/pingcap/ticdc/issues/3487 for release-5.3
https://github.com/pingcap/ticdc/issues/3711 of master branch need a better solution
What is changed and how it works?
we use globalCheckpoint which is a correct-but-early location to reset binlog streamer, and open one transaction safemode
add a half of an integration test back
Check List
Tests
Code changes
Side effects
Related changes
Release note