After Scaling Down TiDB Node and TiKV Node, Status is NA

Note:
This topic has been translated from a Chinese forum by GPT and might contain errors.

Original topic: tidb节点tikv节点缩容后、状态为NA

| username: JOJO_DB

[TiDB Usage Environment] Production Environment / Testing / PoC
[TiDB Version]
[Reproduction Path] What operations were performed that led to the issue
[Encountered Issue: Issue Phenomenon and Impact]


As shown in the picture, 106 is a KV node. After scaling down, its status became Tombstone. Then, after executing pd-ctl store remove-tombstone, it changed to NA status. How can I completely decommission this node?
[Resource Configuration]
[Attachments: Screenshots / Logs / Monitoring]

| username: ffeenn | Original post link

What is the version? If the offline operation is unsuccessful, back up the data!!! Use the force parameter to force offline only as a last resort.
Refer to the following articles:

| username: JOJO_DB | Original post link

Version 5.3.0 has been offline for a long time. Moreover, the physical machine has also been removed. It’s just that the cluster information is still there.

| username: ffeenn | Original post link

First, confirm that there is no data, then back up the data, and use the force parameter to delete it forcibly.

| username: 裤衩儿飞上天 | Original post link

Check if this store still exists using pd-ctl store.

| username: ffeenn | Original post link

There is a high probability that this store exists. This situation arises from the removal of physical data without proper offline procedures.

| username: JOJO_DB | Original post link

No, it has been checked here.

| username: 裤衩儿飞上天 | Original post link

Try deleting this node in the tiup toml file.
Make a backup before deleting.

| username: Running | Original post link

You can force offline using the force parameter.

| username: Jellybean | Original post link

The physical machines are no longer available. You can use the force option of scale in to clean up this type of topology information.

| username: songxuecheng | Original post link

For certain components, the service will not be stopped immediately and the data will not be deleted. Instead, after the data scheduling is completed, the user needs to manually execute the tiup cluster prune command to clean up.

| username: Hacker007 | Original post link

This forced logout will do.

| username: h5n1 | Original post link

After the TiKV status changes to tombstone, you should use tiup cluster prune to clean it up. If there is still residual monitoring information, use pd-ctl store remove-tombstone. Now, you can directly use scale-in -N xxx --force to force delete it.

| username: JOJO_DB | Original post link

After executing scale-in -N --force, the following error is prompted: Error: failed to scale in: cannot find node id ‘192.168.90.106’ in topology. It still cannot be deleted~~

| username: h5n1 | Original post link

Backup the file located at home directory .tiup/storage/cluster/clusters/{cluster-name}/meta.yaml, and then delete the corresponding node content inside.

| username: 考试没答案 | Original post link

Enter the PD console. Check the store limit to see if the node is still there.

| username: 考试没答案 | Original post link

If the node is still there, confirm that the node is not needed and then perform unsafe operations. Try deleting it first. If that doesn’t work, then use unsafe.

| username: JOJO_DB | Original post link

Okay, thank you~ I’ll try it during non-production time.

| username: system | Original post link

This topic was automatically closed 60 days after the last reply. New replies are no longer allowed.