What methods do you use for preloading large amounts of data?

Note:
This topic has been translated from a Chinese forum by GPT and might contain errors.

Original topic: 请问各位都是用什么方式进行大量数据预埋的呢?

| username: 滴滴嗒嘀嗒

To perform performance testing, you need to write a large amount of data into the database first. What methods do you all use? The more, the better!

| username: Miracle | Original post link

Both benchmark and sysbench come with data writing capabilities.

| username: Kongdom | Original post link

You can refer to the official documentation:

| username: Fly-bird | Original post link

Database vendors generally provide testing tools.

| username: Kongdom | Original post link

Indeed, the official documentation provides a testing plan:

| username: 滴滴嗒嘀嗒 | Original post link

Of course, I know the official solution. :smile:

| username: cassblanca | Original post link

Download the TPC-H dataset TPC-H Advanced Sort

| username: gary | Original post link

Both sysbench and tpcc benchmarking tools can generate data.

| username: TiDBer_小阿飞 | Original post link

Write it yourself
mysqlslap -uroot -pHangge_123 --concurrency=10 --number-of-queries=1000000 --create-schema=hangge --query="INSERT INTO people(name, age) VALUES (uuid(), RAND() * 100)"

| username: zhanggame1 | Original post link

Navicat 16, right-click on the table to generate data, the speed is very fast.

| username: 随缘天空 | Original post link

You can use the sysbench tool along with the official scripts. For a data volume of 200-300 million, it takes about 40-50 minutes to import. You can refer to the following link for benchmarking. I have practiced this during my previous benchmarking tests. https://juejin.cn/post/7264921418802462754

| username: system | Original post link

This topic was automatically closed 60 days after the last reply. New replies are no longer allowed.