Migrate Data from SQL Files to TiDB

This document describes how to migrate data from MySQL SQL files to TiDB using TiDB Lightning. For how to generate MySQL SQL files, refer to Export to SQL files using Dumpling.

Prerequisites

Step 1. Prepare SQL files

Put all the SQL files in the same directory, like /data/my_datasource/ or s3://my-bucket/sql-backup. TiDB Lightning recursively searches for all .sql files in this directory and its subdirectories.

Step 2. Define the target table schema

To import data to TiDB, you need to create the table schema for the target database.

If you use Dumpling to export data, the table schema file is automatically exported. For the data exported in other ways, you can create the table schema in one of the following methods:

  • Method 1: Create the target table schema using TiDB Lightning.

    Create SQL files that contain the required DDL statements:

    • Add CREATE DATABASE statements in the ${db_name}-schema-create.sql files.
    • Add CREATE TABLE statements in the ${db_name}.${table_name}-schema.sql files.
  • Method 2: Create the target table schema manually.

Step 3. Create the configuration file

Create a tidb-lightning.toml file with the following content:

[lightning] # Log level = "info" file = "tidb-lightning.log" [tikv-importer] # "local": Default. The local backend is used to import large volumes of data (around or more than 1 TiB). During the import, the target TiDB cluster cannot provide any service. # "tidb": The "tidb" backend can also be used to import small volumes of data (less than 1 TiB). During the import, the target TiDB cluster can provide service normally. For the information about backend mode, refer to https://docs.pingcap.com/tidb/stable/tidb-lightning-backends. backend = "local" # Sets the temporary storage directory for the sorted key-value files. The directory must be empty, and the storage space must be greater than the size of the dataset to be imported. For better import performance, it is recommended to use a directory different from `data-source-dir` and use flash storage and exclusive I/O for the directory. sorted-kv-dir = "${sorted-kv-dir}" [mydumper] # Directory of the data source data-source-dir = "${data-path}" # Local or S3 path, such as 's3://my-bucket/sql-backup' [tidb] # The information of target cluster host = ${host} # For example, 172.16.32.1 port = ${port} # For example, 4000 user = "${user_name}" # For example, "root" password = "${password}" # For example, "rootroot" status-port = ${status-port} # During the import process, TiDB Lightning needs to obtain table schema information from the "Status Port" of TiDB, such as 10080. pd-addr = "${ip}:${port}" # The address of the cluster's PD. TiDB Lightning obtains some information through PD, such as 172.16.31.3:2379. When backend = "local", you must correctly specify status-port and pd-addr. Otherwise, the import will encounter errors.

For more information about the configuration file, refer to TiDB Lightning Configuration.

Step 4. Import the data

To start the import, run tidb-lightning. If you launch the program in the command line, the program might exit because of the SIGHUP signal. In this case, it is recommended to run the program with a nohup or screen tool.

If you import the data from S3, you need to pass in SecretKey and AccessKey of the account as environment variables. The account has the permission to access the S3 backend storage.

export AWS_ACCESS_KEY_ID=${access_key} export AWS_SECRET_ACCESS_KEY=${secret_key} nohup tiup tidb-lightning -config tidb-lightning.toml > nohup.out 2>&1 &

TiDB Lightning also supports reading credential files from ~/.aws/credentials.

After the import is started, you can check the progress in one of the following ways:

After the import is completed, TiDB Lightning automatically exits. Check whether tidb-lightning.log contains the whole procedure completed in the last lines. If yes, the import is successful. If no, the import encounters an error. Address the error as instructed in the error message.