Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

kvclient(ticdc): fix kvclient takes too long time to recover (#3612) #3660

Conversation

ti-chi-bot
Copy link
Member

This is an automated cherry-pick of #3612

close #3191
close flaky test in kvclient: #2694 #3302 #2349 #2688 #2747

What problem does this PR solve?

  1. When a tikv node which has 10k region leaders fail, we found it need more than 30min to recover.
  2. We think it is abnormal for the reason that pd and tikv only need about 30s to tag a node 'disconnect' and elect a new leader.

What is changed and how it works?

  1. Decrease the retry num when a new stream fail to establish to make other region try ASAP.
  2. Remove PartialClone when the region fail.

Reason

  • Expectation
    We use a tikv node which has about 3k region leaders as our comparison object. After test in normal case and abnormal case, we got follow results:
(1) ~ 1min for 3k region to finish init scan and get event (normal case)
(2) > 20min for all 3k region to recover the stream (tikv node fail)

But, pd and tikv only need about 30s to tag a node 'disconnect' and elect a new leader, so a reasonable timespan to recover is about:

       1min30s(normal case + region failover) ~ 5min (retry control in cdc)
  • Consider the probability of network jitter >> tikv temporary down >> tikv permanent down, we still need some retry logic, so
Decrease the retry num when a new stream fail to establish  to make other region try ASAP.
  • When a region fail and call RegionCache.OnRegionFail, it will (1) mark the store as 'needcheck' and fetch new store state of the store asynchronous, (2) move its leader to other peer to try next time and fetch new leader back. This mechanism is suitable in our case, but we find all failed region still connect to the old store which is quite strange. So we
Remove wrong PartialClone logic when the region fail in region_worker.

Result

3min30s to recover from 3k fail (after this pr)  compared to >20min (before) 

Check List

Tests

  • Manual test (add detailed scripts or steps below)

Related changes

  • Need to cherry-pick to the release branch
  • Need to update the documentation
  • Need to update key monitor metrics in both TiCDC document and official document

Release note

Please add a release note.
fix kvclient takes too long time to recover

@ti-chi-bot
Copy link
Member Author

[REVIEW NOTIFICATION]

This pull request has not been approved.

To complete the pull request process, please ask the reviewers in the list to review by filling /cc @reviewer in the comment.
After your PR has acquired the required number of LGTMs, you can assign this pull request to the committer in the list by filling /assign @committer in the comment to help you merge this pull request.

The full list of commands accepted by this bot can be found here.

Reviewer can indicate their review by submitting an approval review.
Reviewer can cancel approval by submitting a request changes review.

@ti-chi-bot ti-chi-bot added release-note Denotes a PR that will be considered when it comes time to generate release notes. do-not-merge/cherry-pick-not-approved labels Nov 29, 2021
@ti-chi-bot
Copy link
Member Author

@ti-chi-bot: This cherry pick PR is for a release branch and has not yet been approved by release team.
Adding the do-not-merge/cherry-pick-not-approved label.

To merge this cherry pick, it must first be approved by the collaborators.

AFTER it has been approved by collaborators, please ping the release team in a comment to request a cherry pick review.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@ti-chi-bot ti-chi-bot added component/kv-client TiKV kv log client component. size/M Denotes a PR that changes 30-99 lines, ignoring generated files. status/LGT2 Indicates that a PR has LGTM 2. type/cherry-pick-for-release-5.0 This PR is cherry-picked to release-5.0 from a source PR. labels Nov 29, 2021
@maxshuang
Copy link
Contributor

/invite

@maxshuang
Copy link
Contributor

/run-kafka-integration-test
/run-integration-test

@overvenus overvenus added this to the v5.0.7 milestone Jan 11, 2022
@ti-chi-bot
Copy link
Member Author

@ti-chi-bot: PR needs rebase.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@ti-chi-bot ti-chi-bot added the needs-rebase Indicates a PR cannot be merged because it has merge conflicts with HEAD. label Jan 20, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
component/kv-client TiKV kv log client component. do-not-merge/cherry-pick-not-approved needs-rebase Indicates a PR cannot be merged because it has merge conflicts with HEAD. release-note Denotes a PR that will be considered when it comes time to generate release notes. size/M Denotes a PR that changes 30-99 lines, ignoring generated files. status/LGT2 Indicates that a PR has LGTM 2. type/cherry-pick-for-release-5.0 This PR is cherry-picked to release-5.0 from a source PR.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants