本节采用单机环境,搭建TiDB测试环境。
全程云环境部署,操作系统为CentOS7.6,用户为root。
1、修改ssh配置
# 提高连接数 vi /etc/ssh/sshd_config MaxSessions 20 #重启sshd service sshd restart
2、安装tidb
# 系统更新 yum -y update # 安装tidb源 curl --proto '=https' --tlsv1.2 -sSf https://tiup-mirrors.pingcap.com/install.sh | sh # 安装tiup cluster source .bash_profile tiup cluster
3、新建cluster配置文件
# 新建配置文件 vi mytidb.yaml # # Global variables are applied to all deployments and used as the default value of # # the deployments if a specific deployment value is missing. global: user: "tidb" ssh_port: 22 deploy_dir: "/tidb-deploy" data_dir: "/tidb-data" # # Monitored variables are applied to all the machines. monitored: node_exporter_port: 9100 blackbox_exporter_port: 9115 server_configs: tidb: log.slow-threshold: 300 tikv: readpool.storage.use-unified-pool: false readpool.coprocessor.use-unified-pool: true pd: replication.enable-placement-rules: true tiflash: logger.level: "info" pd_servers: - host: 192.168.1.111 tidb_servers: - host: 192.168.1.111 tikv_servers: - host: 192.168.1.111 port: 20160 status_port: 20180 - host: 192.168.1.111 port: 20161 status_port: 20181 - host: 192.168.1.111 port: 20162 status_port: 20182 tiflash_servers: - host: 192.168.1.111 monitoring_servers: - host: 192.168.1.111 grafana_servers: - host: 192.168.1.111
4、应用cluster
#应用配置文件 tiup cluster deploy mytidb v4.0.0 ./mytidb.yaml --user root -i hwk8s.pem Starting component `cluster`: /root/.tiup/components/cluster/v1.0.7/tiup-cluster deploy mytidb v4.0.0 ./mytidb.yaml --user root -i hwk8s.pem Please confirm your topology: TiDB Cluster: mytidb TiDB Version: v4.0.0 Type Host Ports OS/Arch Directories ---- ---- ----- ------- ----------- pd 192.168.1.111 2379/2380 linux/x86_64 /tidb-deploy/pd-2379,/tidb-data/pd-2379 tikv 192.168.1.111 20160/20180 linux/x86_64 /tidb-deploy/tikv-20160,/tidb-data/tikv-20160 tikv 192.168.1.111 20161/20181 linux/x86_64 /tidb-deploy/tikv-20161,/tidb-data/tikv-20161 tikv 192.168.1.111 20162/20182 linux/x86_64 /tidb-deploy/tikv-20162,/tidb-data/tikv-20162 tidb 192.168.1.111 4000/10080 linux/x86_64 /tidb-deploy/tidb-4000 tiflash 192.168.1.111 9000/8123/3930/20170/20292/8234 linux/x86_64 /tidb-deploy/tiflash-9000,/tidb-data/tiflash-9000 prometheus 192.168.1.111 9090 linux/x86_64 /tidb-deploy/prometheus-9090,/tidb-data/prometheus-9090 grafana 192.168.1.111 3000 linux/x86_64 /tidb-deploy/grafana-3000 Attention: 1. If the topology is not what you expected, check your yaml file. 2. Please confirm there is no port/directory conflicts in same host. Do you want to continue? [y/N]: y + Generate SSH keys ... Done + Download TiDB components - Download pd:v4.0.0 (linux/amd64) ... Done - Download tikv:v4.0.0 (linux/amd64) ... Done - Download tidb:v4.0.0 (linux/amd64) ... Done - Download tiflash:v4.0.0 (linux/amd64) ... Done - Download prometheus:v4.0.0 (linux/amd64) ... Done - Download grafana:v4.0.0 (linux/amd64) ... Done - Download node_exporter:v0.17.0 (linux/amd64) ... Done - Download blackbox_exporter:v0.12.0 (linux/amd64) ... Done + Initialize target host environments - Prepare 192.168.1.111:22 ... Done + Copy files - Copy pd -> 192.168.1.111 ... Done - Copy tikv -> 192.168.1.111 ... Done - Copy tikv -> 192.168.1.111 ... Done - Copy tikv -> 192.168.1.111 ... Done - Copy tidb -> 192.168.1.111 ... Done - Copy tiflash -> 192.168.1.111 ... Done - Copy prometheus -> 192.168.1.111 ... Done - Copy grafana -> 192.168.1.111 ... Done - Copy node_exporter -> 192.168.1.111 ... Done - Copy blackbox_exporter -> 192.168.1.111 ... Done + Check status Deployed cluster `mytidb` successfully, you can start the cluster via `tiup cluster start mytidb` #启用cluster tiup cluster start mytidb Starting component `cluster`: /root/.tiup/components/cluster/v1.0.7/tiup-cluster start mytidb Starting cluster mytidb... + [ Serial ] - SSHKeySet: privateKey=/root/.tiup/storage/cluster/clusters/mytidb/ssh/id_rsa, publicKey=/root/.tiup/storage/cluster/clusters/mytidb/ssh/id_rsa.pub + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [Parallel] - UserSSH: user=tidb, host=192.168.1.111 + [ Serial ] - ClusterOperate: operation=StartOperation, options={Roles:[] Nodes:[] Force:false SSHTimeout:5 OptTimeout:60 APITimeout:300 IgnoreConfigCheck:false RetainDataRoles:[] RetainDataNodes:[]} Starting component pd Starting instance pd 192.168.1.111:2379 Start pd 192.168.1.111:2379 success Starting component node_exporter Starting instance 192.168.1.111 Start 192.168.1.111 success Starting component blackbox_exporter Starting instance 192.168.1.111 Start 192.168.1.111 success Starting component tikv Starting instance tikv 192.168.1.111:20162 Starting instance tikv 192.168.1.111:20161 Starting instance tikv 192.168.1.111:20160 Start tikv 192.168.1.111:20162 success Start tikv 192.168.1.111:20161 success Start tikv 192.168.1.111:20160 success Starting component tidb Starting instance tidb 192.168.1.111:4000 Start tidb 192.168.1.111:4000 success Starting component tiflash Starting instance tiflash 192.168.1.111:9000 Start tiflash 192.168.1.111:9000 success Starting component prometheus Starting instance prometheus 192.168.1.111:9090 Start prometheus 192.168.1.111:9090 success Starting component grafana Starting instance grafana 192.168.1.111:3000 Start grafana 192.168.1.111:3000 success Checking service state of pd 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:37 CST; 13s ago Checking service state of tikv 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:38 CST; 12s ago 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:38 CST; 12s ago 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:38 CST; 12s ago Checking service state of tidb 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:42 CST; 9s ago Checking service state of tiflash 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:45 CST; 5s ago Checking service state of prometheus 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:47 CST; 4s ago Checking service state of grafana 192.168.1.111 Active: active (running) since Thu 2020-07-02 11:38:47 CST; 4s ago + [ Serial ] - UpdateTopology: cluster=mytidb Started cluster `mytidb` successfully
5、查看cluster状态
#查看cluster清单 tiup cluster list Starting component `cluster`: /root/.tiup/components/cluster/v1.0.7/tiup-cluster list Name User Version Path PrivateKey ---- ---- ------- ---- ---------- mytidb tidb v4.0.0 /root/.tiup/storage/cluster/clusters/mytidb /root/.tiup/storage/cluster/clusters/mytidb/ssh/id_rsa #查看cluster详情 tiup cluster display mytidb Starting component `cluster`: /root/.tiup/components/cluster/v1.0.7/tiup-cluster display mytidb TiDB Cluster: mytidb TiDB Version: v4.0.0 ID Role Host Ports OS/Arch Status Data Dir Deploy Dir -- ---- ---- ----- ------- ------ -------- ---------- 192.168.1.111:3000 grafana 192.168.1.111 3000 linux/x86_64 Up - /tidb-deploy/grafana-3000 192.168.1.111:2379 pd 192.168.1.111 2379/2380 linux/x86_64 Up|L|UI /tidb-data/pd-2379 /tidb-deploy/pd-2379 192.168.1.111:9090 prometheus 192.168.1.111 9090 linux/x86_64 Up /tidb-data/prometheus-9090 /tidb-deploy/prometheus-9090 192.168.1.111:4000 tidb 192.168.1.111 4000/10080 linux/x86_64 Up - /tidb-deploy/tidb-4000 192.168.1.111:9000 tiflash 192.168.1.111 9000/8123/3930/20170/20292/8234 linux/x86_64 Up /tidb-data/tiflash-9000 /tidb-deploy/tiflash-9000 192.168.1.111:20160 tikv 192.168.1.111 20160/20180 linux/x86_64 Up /tidb-data/tikv-20160 /tidb-deploy/tikv-20160 192.168.1.111:20161 tikv 192.168.1.111 20161/20181 linux/x86_64 Up /tidb-data/tikv-20161 /tidb-deploy/tikv-20161 192.168.1.111:20162 tikv 192.168.1.111 20162/20182 linux/x86_64 Up /tidb-data/tikv-20162 /tidb-deploy/tikv-20162
6、mysql客户端操作tidb
#安装源 wget https://repo.mysql.com//mysql80-community-release-el7-3.noarch.rpm rpm -Uvh mysql80-community-release-el7-3.noarch.rpm #安装mysql客户端 yum install mysql-community-client.x86_64 #登录tidb mysql -h 192.168.1.111 -P 4000 -u root #和普通mysql操作区别很小
7、查看tidb管理界面
# 性能监控 http://192.168.1.111:3000 admin/admin # 管理界面 http://192.168.1.111:2379/dashboard root/空