| title | TiDB Environment and System Configuration Check | |
|---|---|---|
| summary | Learn the environment check operations before deploying TiDB. | |
| aliases |
|
This document describes the environment check operations before deploying TiDB. The following steps are ordered by priorities.
For production deployments, it is recommended to use NVMe SSD of EXT4 filesystem to store TiKV data. This configuration is the best practice, whose reliability, security, and stability have been proven in a large number of online scenarios.
Log in to the target machines using the root user account.
Format your data disks to the ext4 filesystem and add the nodelalloc and noatime mount options to the filesystem. It is required to add the nodelalloc option, or else the TiUP deployment cannot pass the precheck. The noatime option is optional.
Note:
If your data disks have been formatted to ext4 and have added the mount options, you can uninstall it by running the
umount /dev/nvme0n1p1command, skip directly to the fifth step below to edit the/etc/fstabfile, and add the options again to the filesystem.
Take the /dev/nvme0n1 data disk as an example:
-
View the data disk.
{{< copyable "shell-root" >}}
fdisk -l
Disk /dev/nvme0n1: 1000 GB -
Create the partition.
{{< copyable "shell-root" >}}
parted -s -a optimal /dev/nvme0n1 mklabel gpt -- mkpart primary ext4 1 -1
Note:
Use the
lsblkcommand to view the device number of the partition: for a NVMe disk, the generated device number is usuallynvme0n1p1; for a regular disk (for example,/dev/sdb), the generated device number is usuallysdb1. -
Format the data disk to the ext4 filesystem.
{{< copyable "shell-root" >}}
mkfs.ext4 /dev/nvme0n1p1
-
View the partition UUID of the data disk.
In this example, the UUID of nvme0n1p1 is
c51eb23b-195c-4061-92a9-3fad812cc12f.{{< copyable "shell-root" >}}
lsblk -f
NAME FSTYPE LABEL UUID MOUNTPOINT sda ├─sda1 ext4 237b634b-a565-477b-8371-6dff0c41f5ab /boot ├─sda2 swap f414c5c0-f823-4bb1-8fdf-e531173a72ed └─sda3 ext4 547909c1-398d-4696-94c6-03e43e317b60 / sr0 nvme0n1 └─nvme0n1p1 ext4 c51eb23b-195c-4061-92a9-3fad812cc12f -
Edit the
/etc/fstabfile and add thenodelallocmount options.{{< copyable "shell-root" >}}
vi /etc/fstab
UUID=c51eb23b-195c-4061-92a9-3fad812cc12f /data1 ext4 defaults,nodelalloc,noatime 0 2 -
Mount the data disk.
{{< copyable "shell-root" >}}
mkdir /data1 && \ mount -a -
Check using the following command.
{{< copyable "shell-root" >}}
mount -t ext4
/dev/nvme0n1p1 on /data1 type ext4 (rw,noatime,nodelalloc,data=ordered)If the filesystem is ext4 and
nodelallocis included in the mount options, you have successfully mount the data disk ext4 filesystem with options on the target machines.
TiDB needs sufficient memory space for operation. When memory is insufficient, using swap as a buffer might degrade performance. Therefore, it is recommended to disable the system swap permanently by executing the following commands:
{{< copyable "shell-regular" >}}
echo "vm.swappiness = 0">> /etc/sysctl.conf
swapoff -a && swapon -a
sysctl -pNote:
Executing
swapoff -aand thenswapon -ais to refresh swap by dumping data to memory and cleaning up swap. If you drop the swappiness change and execute onlyswapoff -a, swap will be enabled again after you restart the system.
sysctl -pis to make the configuration effective without restarting the system.
In TiDB clusters, the access ports between nodes must be open to ensure the transmission of information such as read and write requests and data heartbeats. In common online scenarios, the data interaction between the database and the application service and between the database nodes are all made within a secure network. Therefore, if there are no special security requirements, it is recommended to stop the firewall of the target machine. Otherwise, refer to the port usage and add the needed port information to the allowlist of the firewall service.
The rest of this section describes how to stop the firewall service of a target machine.
-
Check the firewall status. Take CentOS Linux release 7.7.1908 (Core) as an example.
{{< copyable "shell-regular" >}}
sudo firewall-cmd --state sudo systemctl status firewalld.service
-
Stop the firewall service.
{{< copyable "shell-regular" >}}
sudo systemctl stop firewalld.service
-
Disable automatic start of the firewall service.
{{< copyable "shell-regular" >}}
sudo systemctl disable firewalld.service
-
Check the firewall status.
{{< copyable "shell-regular" >}}
sudo systemctl status firewalld.service
TiDB is a distributed database system that requires clock synchronization between nodes to guarantee linear consistency of transactions in the ACID model.
At present, the common solution to clock synchronization is to use the Network Time Protocol (NTP) services. You can use the pool.ntp.org timing service on the Internet, or build your own NTP service in an offline environment.
To check whether the NTP service is installed and whether it synchronizes with the NTP server normally, take the following steps:
-
Run the following command. If it returns
running, then the NTP service is running.{{< copyable "shell-regular" >}}
sudo systemctl status ntpd.service
ntpd.service - Network Time Service Loaded: loaded (/usr/lib/systemd/system/ntpd.service; disabled; vendor preset: disabled) Active: active (running) since 一 2017-12-18 13:13:19 CST; 3s ago-
If it returns
Unit ntpd.service could not be found., then try the following command to see whether your system is configured to usechronydinstead ofntpdto perform clock synchronization with NTP:{{< copyable "shell-regular" >}}
sudo systemctl status chronyd.service
chronyd.service - NTP client/server Loaded: loaded (/usr/lib/systemd/system/chronyd.service; enabled; vendor preset: enabled) Active: active (running) since Mon 2021-04-05 09:55:29 EDT; 3 days agoIf the result shows that neither
chronydnorntpdis configured, it means that neither of them is installed in your system. You should first installchronydorntpdand ensure that it can be automatically started. By default,ntpdis used.If your system is configured to use
chronyd, proceed to step 3.
-
-
Run the
ntpstatcommand to check whether the NTP service synchronizes with the NTP server.Note:
For the Ubuntu system, you need to install the
ntpstatpackage.{{< copyable "shell-regular" >}}
ntpstat
-
If it returns
synchronised to NTP server(synchronizing with the NTP server), then the synchronization process is normal.synchronised to NTP server (85.199.214.101) at stratum 2 time correct to within 91 ms polling server every 1024 s -
The following situation indicates the NTP service is not synchronizing normally:
unsynchronised -
The following situation indicates the NTP service is not running normally:
Unable to talk to NTP daemon. Is it running?
-
-
Run the
chronyc trackingcommand to check wheter the Chrony service synchronizes with the NTP server.Note:
This only applies to systems that use Chrony instead of NTPd.
{{< copyable "shell-regular" >}}
chronyc tracking
-
If the command returns
Leap status : Normal, the synchronization process is normal.Reference ID : 5EC69F0A (ntp1.time.nl) Stratum : 2 Ref time (UTC) : Thu May 20 15:19:08 2021 System time : 0.000022151 seconds slow of NTP time Last offset : -0.000041040 seconds RMS offset : 0.000053422 seconds Frequency : 2.286 ppm slow Residual freq : -0.000 ppm Skew : 0.012 ppm Root delay : 0.012706812 seconds Root dispersion : 0.000430042 seconds Update interval : 1029.8 seconds Leap status : Normal -
If the command returns the following result, an error occurs in the synchronization:
Leap status : Not synchronised -
If the command returns the following result, the
chronydservice is not running normally:506 Cannot talk to daemon
-
To make the NTP service start synchronizing as soon as possible, run the following command. Replace pool.ntp.org with your NTP server.
{{< copyable "shell-regular" >}}
sudo systemctl stop ntpd.service && \
sudo ntpdate pool.ntp.org && \
sudo systemctl start ntpd.serviceTo install the NTP service manually on the CentOS 7 system, run the following command:
{{< copyable "shell-regular" >}}
sudo yum install ntp ntpdate && \
sudo systemctl start ntpd.service && \
sudo systemctl enable ntpd.serviceFor TiDB in the production environment, it is recommended to optimize the operating system configuration in the following ways:
- Disable THP (Transparent Huge Pages). The memory access pattern of databases tends to be sparse rather than consecutive. If the high-level memory fragmentation is serious, higher latency will occur when THP pages are allocated.
- Set the I/O Scheduler of the storage media to
noop. For the high-speed SSD storage media, the kernel's I/O scheduling operations can cause performance loss. After the Scheduler is set tonoop, the performance is better because the kernel directly sends I/O requests to the hardware without other operations. Also, the noop Scheduler is better applicable. - Choose the
performancemode for the cpufrequ module which controls the CPU frequency. The performance is maximized when the CPU frequency is fixed at its highest supported operating frequency without dynamic adjustment.
Take the following steps to check the current operating system configuration and configure optimal parameters:
-
Execute the following command to see whether THP is enabled or disabled:
{{< copyable "shell-regular" >}}
cat /sys/kernel/mm/transparent_hugepage/enabled
[always] madvise neverNote:
If
[always] madvise neveris output, THP is enabled. You need to disable it. -
Execute the following command to see the I/O Scheduler of the disk where the data directory is located. Assume that you create data directories on both sdb and sdc disks:
{{< copyable "shell-regular" >}}
cat /sys/block/sd[bc]/queue/scheduler
noop [deadline] cfq noop [deadline] cfqNote:
If
noop [deadline] cfqis output, the I/O Scheduler for the disk is in thedeadlinemode. You need to change it tonoop. -
Execute the following command to see the
ID_SERIALof the disk:{{< copyable "shell-regular" >}}
udevadm info --name=/dev/sdb | grep ID_SERIALE: ID_SERIAL=36d0946606d79f90025f3e09a0c1f9e81 E: ID_SERIAL_SHORT=6d0946606d79f90025f3e09a0c1f9e81Note:
If multiple disks are allocated with data directories, you need to execute the above command several times to record the
ID_SERIALof each disk. -
Execute the following command to see the power policy of the cpufreq module:
{{< copyable "shell-regular" >}}
cpupower frequency-info --policy
analyzing CPU 0: current policy: frequency should be within 1.20 GHz and 3.10 GHz. The governor "powersave" may decide which speed to use within this range.Note:
If
The governor "powersave"is output, the power policy of the cpufreq module ispowersave. You need to modify it toperformance. If you use a virtual machine or a cloud host, the output is usuallyUnable to determine current policy, and you do not need to change anything. -
Configure optimal parameters of the operating system:
-
Method one: Use tuned (Recommended)
-
Execute the
tuned-adm listcommand to see the tuned profile of the current operating system:{{< copyable "shell-regular" >}}
tuned-adm list
Available profiles: - balanced - General non-specialized tuned profile - desktop - Optimize for the desktop use-case - hpc-compute - Optimize for HPC compute workloads - latency-performance - Optimize for deterministic performance at the cost of increased power consumption - network-latency - Optimize for deterministic performance at the cost of increased power consumption, focused on low latency network performance - network-throughput - Optimize for streaming network throughput, generally only necessary on older CPUs or 40G+ networks - powersave - Optimize for low power consumption - throughput-performance - Broadly applicable tuning that provides excellent performance across a variety of common server workloads - virtual-guest - Optimize for running inside a virtual guest - virtual-host - Optimize for running KVM guests Current active profile: balancedThe output
Current active profile: balancedmeans that the tuned profile of the current operating system isbalanced. It is recommended to optimize the configuration of the operating system based on the current profile. -
Create a new tuned profile:
{{< copyable "shell-regular" >}}
mkdir /etc/tuned/balanced-tidb-optimal/ vi /etc/tuned/balanced-tidb-optimal/tuned.conf
[main] include=balanced [cpu] governor=performance [vm] transparent_hugepages=never [disk] devices_udev_regex=(ID_SERIAL=36d0946606d79f90025f3e09a0c1fc035)|(ID_SERIAL=36d0946606d79f90025f3e09a0c1f9e81) elevator=noopThe output
include=balancedmeans to add the optimization configuration of the operating system to the currentbalancedprofile. -
Apply the new tuned profile:
{{< copyable "shell-regular" >}}
tuned-adm profile balanced-tidb-optimal
-
-
Method two: Configure using scripts. Skip this method if you already use method one.
-
Execute the
grubbycommand to see the default kernel version:Note:
Install the
grubbypackage first before you executegrubby.{{< copyable "shell-regular" >}}
grubby --default-kernel
/boot/vmlinuz-3.10.0-957.el7.x86_64
-
Execute
grubby --update-kernelto modify the kernel configuration:{{< copyable "shell-regular" >}}
grubby --args="transparent_hugepage=never" --update-kernel /boot/vmlinuz-3.10.0-957.el7.x86_64Note:
--update-kernelis followed by the actual default kernel version. -
Execute
grubby --infoto see the modified default kernel configuration:{{< copyable "shell-regular" >}}
grubby --info /boot/vmlinuz-3.10.0-957.el7.x86_64
Note:
--infois followed by the actual default kernel version.index=0 kernel=/boot/vmlinuz-3.10.0-957.el7.x86_64 args="ro crashkernel=auto rd.lvm.lv=centos/root rd.lvm.lv=centos/swap rhgb quiet LANG=en_US.UTF-8 transparent_hugepage=never" root=/dev/mapper/centos-root initrd=/boot/initramfs-3.10.0-957.el7.x86_64.img title=CentOS Linux (3.10.0-957.el7.x86_64) 7 (Core) -
Modify the current kernel configuration to immediately disable THP:
{{< copyable "shell-regular" >}}
echo never > /sys/kernel/mm/transparent_hugepage/enabled echo never > /sys/kernel/mm/transparent_hugepage/defrag
-
Configure the I/O Scheduler in the udev script:
{{< copyable "shell-regular" >}}
vi /etc/udev/rules.d/60-tidb-schedulers.rules
ACTION=="add|change", SUBSYSTEM=="block", ENV{ID_SERIAL}=="36d0946606d79f90025f3e09a0c1fc035", ATTR{queue/scheduler}="noop" ACTION=="add|change", SUBSYSTEM=="block", ENV{ID_SERIAL}=="36d0946606d79f90025f3e09a0c1f9e81", ATTR{queue/scheduler}="noop" -
Apply the udev script:
{{< copyable "shell-regular" >}}
udevadm control --reload-rules udevadm trigger --type=devices --action=change
-
Create a service to configure the CPU power policy:
{{< copyable "shell-regular" >}}
cat >> /etc/systemd/system/cpupower.service << EOF [Unit] Description=CPU performance [Service] Type=oneshot ExecStart=/usr/bin/cpupower frequency-set --governor performance [Install] WantedBy=multi-user.target EOF
-
Apply the CPU power policy configuration service:
{{< copyable "shell-regular" >}}
systemctl daemon-reload systemctl enable cpupower.service systemctl start cpupower.service
-
-
-
Execute the following command to verify the THP status:
{{< copyable "shell-regular" >}}
cat /sys/kernel/mm/transparent_hugepage/enabled
always madvise [never] -
Execute the following command to verify the I/O Scheduler of the disk where the data directory is located:
{{< copyable "shell-regular" >}}
cat /sys/block/sd[bc]/queue/scheduler
[noop] deadline cfq [noop] deadline cfq -
Execute the following command to see the power policy of the cpufreq module:
{{< copyable "shell-regular" >}}
cpupower frequency-info --policy
analyzing CPU 0: current policy: frequency should be within 1.20 GHz and 3.10 GHz. The governor "performance" may decide which speed to use within this range. -
Execute the following commands to modify the
sysctlparameters:{{< copyable "shell-regular" >}}
echo "fs.file-max = 1000000">> /etc/sysctl.conf echo "net.core.somaxconn = 32768">> /etc/sysctl.conf echo "net.ipv4.tcp_tw_recycle = 0">> /etc/sysctl.conf echo "net.ipv4.tcp_syncookies = 0">> /etc/sysctl.conf echo "vm.overcommit_memory = 1">> /etc/sysctl.conf sysctl -p
-
Execute the following command to configure the user's
limits.conffile:{{< copyable "shell-regular" >}}
cat << EOF >>/etc/security/limits.conf tidb soft nofile 1000000 tidb hard nofile 1000000 tidb soft stack 32768 tidb hard stack 32768 EOF
This section describes how to manually configure the SSH mutual trust and sudo without password. It is recommended to use TiUP for deployment, which automatically configure SSH mutual trust and login without password. If you deploy TiDB clusters using TiUP, ignore this section.
-
Log in to the target machine respectively using the
rootuser account, create thetidbuser and set the login password.{{< copyable "shell-root" >}}
useradd tidb && \ passwd tidb -
To configure sudo without password, run the following command, and add
tidb ALL=(ALL) NOPASSWD: ALLto the end of the file:{{< copyable "shell-root" >}}
visudo
tidb ALL=(ALL) NOPASSWD: ALL -
Use the
tidbuser to log in to the control machine, and run the following command. Replace10.0.1.1with the IP of your target machine, and enter thetidbuser password of the target machine as prompted. After the command is executed, SSH mutual trust is already created. This applies to other machines as well. Newly createdtidbusers do not have the.sshdirectory. To create such a directory, execute the command that generates the RSA key. To deploy TiDB components on the control machine, configure mutual trust for the control machine and the control machine itself.{{< copyable "shell-regular" >}}
ssh-keygen -t rsa ssh-copy-id -i ~/.ssh/id_rsa.pub 10.0.1.1 -
Log in to the control machine using the
tidbuser account, and log in to the IP of the target machine usingssh. If you do not need to enter the password and can successfully log in, then the SSH mutual trust is successfully configured.{{< copyable "shell-regular" >}}
ssh 10.0.1.1
[tidb@10.0.1.1 ~]$ -
After you log in to the target machine using the
tidbuser, run the following command. If you do not need to enter the password and can switch to therootuser, then sudo without password of thetidbuser is successfully configured.{{< copyable "shell-regular" >}}
sudo -su root
[root@10.0.1.1 tidb]#
This section describes how to install the NUMA tool. In online environments, because the hardware configuration is usually higher than required, to better plan the hardware resources, multiple instances of TiDB or TiKV can be deployed on a single machine. In such scenarios, you can use NUMA tools to prevent the competition for CPU resources which might cause reduced performance.
Note:
- Binding cores using NUMA is a method to isolate CPU resources and is suitable for deploying multiple instances on highly configured physical machines.
- After completing deployment using
tiup cluster deploy, you can use theexeccommand to perform cluster level management operations.
-
Log in to the target node to install. Take CentOS Linux release 7.7.1908 (Core) as an example.
{{< copyable "shell-regular" >}}
sudo yum -y install numactl
-
Run the
execcommand usingtiup clusterto install in batches.{{< copyable "shell-regular" >}}
tiup cluster exec --helpRun shell command on host in the tidb cluster Usage: cluster exec <cluster-name> [flags] Flags: --command string the command run on cluster host (default "ls") -h, --help help for exec --sudo use root permissions (default false)To use the sudo privilege to execute the installation command for all the target machines in the
tidb-testcluster, run the following command:{{< copyable "shell-regular" >}}
tiup cluster exec tidb-test --sudo --command "yum -y install numactl"