Under the same traffic, using batch read/write in TiKV will increase IO utilization by more than 50%

Note:
This topic has been translated from a Chinese forum by GPT and might contain errors.

Original topic: tikv 在同等的流量下,使用 batch读写, IO utilization 会升高 50% 以上

| username: Doslin

[TiDB Usage Environment] Production Environment
[TiDB Version] v6.1
[Reproduction Path] Operations performed that led to the issue
[Encountered Issue: Issue Phenomenon and Impact]
At 13:23 in the image, I changed raw_put to raw_batch_put and raw_get to raw_batch_get, with a batch size of 60. After making these changes, I noticed that the machine’s IO utilization increased by 80%.

You can see that the traffic before and after the change did not vary much, and even decreased.


Moreover, the metrics for raw_batch_get and raw_batch_put only slightly decreased, but according to my service-side statistics, it is indeed 1/60.

[Resource Configuration]

server_configs:
  tidb:
    log.slow-threshold: 300
  tikv:
    readpool.coprocessor.use-unified-pool: true
    readpool.storage.use-unified-pool: false
  pd:
    replication.enable-placement-rules: true
    replication.location-labels:
    - host
  tidb_dashboard: {}
  tiflash:
    logger.level: info
  tiflash-learner: {}
  pump: {}
  drainer: {}
  cdc: {}
  kvcdc: {}
  grafana: {}

[Attachments: Screenshots/Logs/Monitoring]

| username: 裤衩儿飞上天 | Original post link

Has the read and write speed also improved significantly?

| username: Doslin | Original post link

There was no improvement, but rather a decrease.

| username: Doslin | Original post link

Are there any other troubleshooting ideas?

| username: Min_Chen | Original post link

Hello, please provide the monitoring data. Export it using Clinic, referring to the method in Using PingCAP Clinic to Diagnose the Cluster.

| username: system | Original post link

This topic was automatically closed 60 days after the last reply. New replies are no longer allowed.