As an open source distributed NewSQL database with high performance, TiDB can be deployed in the Intel architecture server, ARM architecture server, and major virtualization environments and runs well. TiDB supports most of the major hardware networks and Linux operating systems.
|Red Hat Enterprise Linux||7.3 or later 7.x releases|
|CentOS||7.3 or later 7.x releases|
|Oracle Enterprise Linux||7.3 or later 7.x releases|
|Ubuntu LTS||16.04 or later|
Other Linux OS versions such as Debian Linux and Fedora Linux might work but are not officially supported.
|sshpass||1.06 or later|
|TiUP||1.5.0 or later|
|sshpass||1.06 or later|
|numa||2.0.12 or later|
You can deploy and run TiDB on the 64-bit generic hardware server platform in the Intel x86-64 architecture or on the hardware server platform in the ARM architecture. The requirements and recommendations about server hardware configuration (ignoring the resources occupied by the operating system itself) for development, test, and production environments are as follows:
|Component||CPU||Memory||Local Storage||Network||Instance Number (Minimum Requirement)|
|TiDB||8 core+||16 GB+||No special requirements||Gigabit network card||1 (can be deployed on the same machine with PD)|
|PD||4 core+||8 GB+||SAS, 200 GB+||Gigabit network card||1 (can be deployed on the same machine with TiDB)|
|TiKV||8 core+||32 GB+||SAS, 200 GB+||Gigabit network card||3|
|TiFlash||32 core+||64 GB+||SSD, 200 GB+||Gigabit network card||1|
|TiCDC||8 core+||16 GB+||SAS, 200 GB+||Gigabit network card||1|
|Component||CPU||Memory||Hard Disk Type||Network||Instance Number (Minimum Requirement)|
|TiDB||16 core+||48 GB+||SAS||10 Gigabit network card (2 preferred)||2|
|PD||8 core+||16 GB+||SSD||10 Gigabit network card (2 preferred)||3|
|TiKV||16 core+||64 GB+||SSD||10 Gigabit network card (2 preferred)||3|
|TiFlash||48 core+||128 GB+||1 or more SSDs||10 Gigabit network card (2 preferred)||2|
|TiCDC||16 core+||64 GB+||SSD||10 Gigabit network card (2 preferred)||2|
|Monitor||8 core+||16 GB+||SAS||Gigabit network card||1|
Before you deploy TiFlash, note the following items:
- TiFlash can be deployed on multiple disks.
- It is recommended to use a high-performance SSD as the first disk of the TiFlash data directory to buffer the real-time replication of TiKV data. The performance of this disk should not be lower than that of TiKV, such as PCI-E SSD. The disk capacity should be no less than 10% of the total capacity; otherwise, it might become the bottleneck of this node. You can deploy ordinary SSDs for other disks, but note that a better PCI-E SSD brings better performance.
- It is recommended to deploy TiFlash on different nodes from TiKV. If you must deploy TiFlash and TiKV on the same node, increase the number of CPU cores and memory, and try to deploy TiFlash and TiKV on different disks to avoid interfering each other.
- The total capacity of the TiFlash disks is calculated in this way:
the data volume of the entire TiKV cluster to be replicated / the number of TiKV replicas * the number of TiFlash replicas. For example, if the overall planned capacity of TiKV is 1 TB, the number of TiKV replicas is 3, and the number of TiFlash replicas is 2, then the recommended total capacity of TiFlash is
1024 GB / 3 * 2. You can replicate only the data of some tables. In such case, determine the TiFlash capacity according to the data volume of the tables to be replicated.
Before you deploy TiCDC, note that it is recommended to deploy TiCDC on PCIe-SSD disks larger than 1 TB.
As an open source distributed NewSQL database, TiDB requires the following network port configuration to run. Based on the TiDB deployment in actual environments, the administrator can open relevant ports in the network side and host side.
|TiDB||4000||the communication port for the application and DBA tools|
|TiDB||10080||the communication port to report TiDB status|
|TiKV||20160||the TiKV communication port|
|TiKV||20180||the communication port to report TiKV status|
|PD||2379||the communication port between TiDB and PD|
|PD||2380||the inter-node communication port within the PD cluster|
|TiFlash||9000||the TiFlash TCP service port|
|TiFlash||8123||the TiFlash HTTP service port|
|TiFlash||3930||the TiFlash RAFT and Coprocessor service port|
|TiFlash||20170||the TiFlash Proxy service port|
|TiFlash||20292||the port for Prometheus to pull TiFlash Proxy metrics|
|TiFlash||8234||the port for Prometheus to pull TiFlash metrics|
|Pump||8250||the Pump communication port|
|Drainer||8249||the Drainer communication port|
|TiCDC||8300||the TiCDC communication port|
|Monitoring||9090||the communication port for the Prometheus service|
|Monitoring||12020||the communication port for the NgMonitoring service|
|Node_exporter||9100||the communication port to report the system information of every TiDB cluster node|
|Blackbox_exporter||9115||the Blackbox_exporter communication port, used to monitor the ports in the TiDB cluster|
|Grafana||3000||the port for the external Web monitoring service and client (Browser) access|
|Alertmanager||9093||the port for the alert web service|
|Alertmanager||9094||the alert communication port|
|Component||Disk space requirement||Healthy disk usage|
|TiDB||At least 30 GB for the log disk||Lower than 90%|
|PD||At least 20 GB for the data disk and for the log disk, respectively||Lower than 90%|
|TiKV||At least 100 GB for the data disk and for the log disk, respectively||Lower than 80%|
|TiFlash||At least 100 GB for the data disk and at least 30 GB for the log disk, respectively||Lower than 80%|