-
Notifications
You must be signed in to change notification settings - Fork 289
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The CDC client is still using the old PD address #9584
Comments
/assign @asddongmen |
@jacktd9 May I ask if the changefeed has resumed normal synchronization? In other words, was the error log you found temporary or has it not been resolved yet? |
So, if I understand correctly, TiCDC's pdClient is still using the old address and cannot update to the new one? Based on our discussion and my comprehension, here is a summary:
Please correct me if I misunderstood anything. @jacktd9 |
Fixed in v6.5.6. |
We have found that it is still not able to create changefeeds after scale-out -> scale-in operation. Reproduction: tiup playground v7.1.5 --db 1 --kv 1 --pd 3 --ticdc 3 --tiflash 0 --without-monitor
# sanity check, ensure that tidb & cdc works
mysql -u root -h 127.1 -P 4000 -e 'select * from mysql.tidb'
tiup cdc:v7.1.5 cli --server 127.0.0.1:8300 changefeed create --sink-uri 'blackhole://' -c test0
tiup cdc:v7.1.5 cli --server 127.0.0.1:8300 changefeed remove -c test0
# perform scale-out (do not scale out all 3 simultaneously!)
tiup playground scale-out --pd 1
tiup playground scale-out --pd 1
tiup playground scale-out --pd 1
# note the PIDs of the first 3 PDs
tiup playground display
# perform scale-in
tiup playground scale-in --pid 23397,23398,23399
# sanity check, tidb should still work
mysql -u root -h 127.1 -P 4000 -e 'select * from mysql.tidb'
mysql -u root -h 127.1 -P 4000 -e 'show config where type = "pd" and name = "cluster-version"'
# try to create changefeed again
tiup cdc:v7.1.5 cli --server 127.0.0.1:8300 changefeed create --sink-uri 'blackhole://' -c test1
# ^ the program above is now stuck.
# cdc log shows a lot of warnings like:
#
# [2025/01/09 15:24:37.834 +08:00] [WARN] [pd_service_discovery.go:370] ["[pd] failed to get cluster id"] [url=http://127.0.0.1:2382] [error="[PD:client:ErrClientGetMember]error:rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 127.0.0.1:2382: connect: connection refused\" target:127.0.0.1:2382 status:TRANSIENT_FAILURE: error:rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 127.0.0.1:2382: connect: connection refused\" target:127.0.0.1:2382 status:TRANSIENT_FAILURE"]
#
# doing Ctrl+C may also give us
#
# Error: [CDC:ErrNewStore]new store failed: [pd] failed to get cluster id This is reproduced on v7.1.4, v7.1.6, v8.5.0. I'm going to reopen the issue. |
@lidezhu No the current behavior can be explained saying the CDC owner handles the API by using the |
It should be a problem of pd: tikv/pd#8993 |
Create a new issue #12004 for the convenience of later cherry pick's management. Close this one. |
What did you do?
An error occurred during the command execution, and it was found that the CDC logs are still accessing the old PD

current PD address is... 2479

What did you expect to see?
no error
What did you see instead?
connect pd failed
Versions of the cluster
cluster version : v6.5.3
The text was updated successfully, but these errors were encountered: