Then the TiKV log keeps printing this:
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
[2023/03/14 09:04:46.461 +08:00] [WARN] [pd.rs:1693] [“failed to update max timestamp for region 63021: Pd(Other("[components/pd_client/src/tso.rs:97]: TimestampRequest channel is closed"))”]
When starting up, this error is reported:
Error: failed to start tikv: failed to start: 10.0.124.78 tikv-20160.service, please check the instance’s log(/tidb-deploy/tikv-20160/log) for more detail.: timed out waiting for port 20160 to be started after 2m0s
The cluster startup operation will start all components of the entire TiDB cluster in the order of PD → TiKV → Pump → TiDB → TiFlash → Drainer → TiCDC → Prometheus → Grafana → Alertmanager. You need to ensure that your PD nodes can start normally before your TiKV nodes can start.
[2023/03/13 20:06:09.158 +08:00] [ERROR] [util.rs:419] [“request failed, retry”] [err_code=KV:PD:gRPC] [err=“Grpc(RpcFailure(RpcStatus { code: 2-UNKNOWN, message: “rpc error: code = Unavailable desc = not leader”, details: }))”] —> This indicates that TiKV sent an RPC message to PD, but it reported an UNKNOWN error. Firstly, it can be confirmed that there is an issue with the TiKV connection to PD. Secondly, this UNKNOWN error feels like the PD or TiKV version hasn’t been upgraded. Confirm if all binaries are up to date (./pd-server --version or ./tikv-server --version).
Having only one PD node that cannot start properly is quite dangerous for production! According to tidb菜鸟一只, please provide a complete pd.log to confirm that PD is running normally and to understand why it was “working fine last Friday but failed on Monday”.
The current information is still not enough to solve this problem.
If this cannot be checked, simply stop checking. Delete this PD from the cluster and start a brand new PD to join the cluster. Ensure the PD is functioning normally first.