From 15c9d956d3eb588cc26630b7bd4091ce43041cfd Mon Sep 17 00:00:00 2001 From: Ran <huangran.alex@gmail.com> Date: Fri, 5 Jan 2024 17:08:47 +0800 Subject: [PATCH 1/4] Add temp.md --- temp.md | 1 + 1 file changed, 1 insertion(+) create mode 100644 temp.md diff --git a/temp.md b/temp.md new file mode 100644 index 0000000000000..af27ff4986a7b --- /dev/null +++ b/temp.md @@ -0,0 +1 @@ +This is a test file. \ No newline at end of file From f2f989f05aa9469b99bb6256fef1c2fdb653fb9c Mon Sep 17 00:00:00 2001 From: Ran <huangran.alex@gmail.com> Date: Fri, 5 Jan 2024 17:08:51 +0800 Subject: [PATCH 2/4] Delete temp.md --- temp.md | 1 - 1 file changed, 1 deletion(-) delete mode 100644 temp.md diff --git a/temp.md b/temp.md deleted file mode 100644 index af27ff4986a7b..0000000000000 --- a/temp.md +++ /dev/null @@ -1 +0,0 @@ -This is a test file. \ No newline at end of file From 47b62b8e1e88d7b06339fb8a317eb31370bb1ae1 Mon Sep 17 00:00:00 2001 From: Ran <huangran.alex@gmail.com> Date: Fri, 5 Jan 2024 17:15:09 +0800 Subject: [PATCH 3/4] bump versions --- br/backup-and-restore-use-cases.md | 6 ++--- dm/maintain-dm-using-tiup.md | 2 +- dm/quick-start-create-task.md | 2 +- pd-control.md | 2 +- production-deployment-using-tiup.md | 4 +-- quick-start-with-tidb.md | 10 +++---- scale-tidb-using-tiup.md | 2 +- ticdc/deploy-ticdc.md | 4 +-- ticdc/ticdc-changefeed-config.md | 6 ++--- ticdc/ticdc-changefeed-overview.md | 2 +- ticdc/ticdc-manage-changefeed.md | 2 +- ticdc/ticdc-sink-to-cloud-storage.md | 2 +- tidb-binlog/get-started-with-tidb-binlog.md | 2 +- tiflash/create-tiflash-replicas.md | 2 +- tiup/tiup-cluster.md | 30 ++++++++++----------- tiup/tiup-component-cluster-deploy.md | 2 +- tiup/tiup-component-cluster-patch.md | 2 +- tiup/tiup-component-cluster-upgrade.md | 2 +- tiup/tiup-component-dm-upgrade.md | 2 +- tiup/tiup-component-management.md | 12 ++++----- tiup/tiup-mirror.md | 6 ++--- upgrade-tidb-using-tiup.md | 25 ++++++++--------- 22 files changed, 65 insertions(+), 64 deletions(-) diff --git a/br/backup-and-restore-use-cases.md b/br/backup-and-restore-use-cases.md index a681b2b1b0b9e..5d403ac2db487 100644 --- a/br/backup-and-restore-use-cases.md +++ b/br/backup-and-restore-use-cases.md @@ -17,7 +17,7 @@ With PITR, you can satisfy the preceding requirements. ## Deploy the TiDB cluster and BR -To use PITR, you need to deploy a TiDB cluster >= v6.2.0 and update BR to the same version as the TiDB cluster. This document uses v6.5.6 as an example. +To use PITR, you need to deploy a TiDB cluster >= v6.2.0 and update BR to the same version as the TiDB cluster. This document uses v6.5.7 as an example. The following table shows the recommended hardware resources for using PITR in a TiDB cluster. @@ -44,13 +44,13 @@ Install or upgrade BR using TiUP: - Install: ```shell - tiup install br:v6.5.6 + tiup install br:v6.5.7 ``` - Upgrade: ```shell - tiup update br:v6.5.6 + tiup update br:v6.5.7 ``` ## Configure backup storage (Amazon S3) diff --git a/dm/maintain-dm-using-tiup.md b/dm/maintain-dm-using-tiup.md index 402e3ead64ffd..0fe8b6f8d5fc5 100644 --- a/dm/maintain-dm-using-tiup.md +++ b/dm/maintain-dm-using-tiup.md @@ -389,7 +389,7 @@ All operations above performed on the cluster machine use the SSH client embedde Then you can use the `--native-ssh` command-line flag to enable the system-native command-line tool: -- Deploy a cluster: `tiup dm deploy <cluster-name> <version> <topo> --native-ssh`. Fill in the name of your cluster for `<cluster-name>`, the DM version to be deployed (such as `v6.5.6`) for `<version>` , and the topology file name for `<topo>`. +- Deploy a cluster: `tiup dm deploy <cluster-name> <version> <topo> --native-ssh`. Fill in the name of your cluster for `<cluster-name>`, the DM version to be deployed (such as `v6.5.7`) for `<version>` , and the topology file name for `<topo>`. - Start a cluster: `tiup dm start <cluster-name> --native-ssh`. - Upgrade a cluster: `tiup dm upgrade ... --native-ssh` diff --git a/dm/quick-start-create-task.md b/dm/quick-start-create-task.md index ebb0ed8fb897c..6d23444a6c1a8 100644 --- a/dm/quick-start-create-task.md +++ b/dm/quick-start-create-task.md @@ -73,7 +73,7 @@ To run a TiDB server, use the following command: {{< copyable "shell-regular" >}} ```bash -wget https://download.pingcap.org/tidb-community-server-v6.5.6-linux-amd64.tar.gz +wget https://download.pingcap.org/tidb-community-server-v6.5.7-linux-amd64.tar.gz tar -xzvf tidb-latest-linux-amd64.tar.gz mv tidb-latest-linux-amd64/bin/tidb-server ./ ./tidb-server diff --git a/pd-control.md b/pd-control.md index 5cb5d487445e6..33b93cc1a0b35 100644 --- a/pd-control.md +++ b/pd-control.md @@ -28,7 +28,7 @@ To obtain `pd-ctl` of the latest version, download the TiDB server installation > **Note:** > -> `{version}` in the link indicates the version number of TiDB. For example, the download link for `v6.5.6` in the `amd64` architecture is `https://download.pingcap.org/tidb-community-server-v6.5.6-linux-amd64.tar.gz`. +> `{version}` in the link indicates the version number of TiDB. For example, the download link for `v6.5.7` in the `amd64` architecture is `https://download.pingcap.org/tidb-community-server-v6.5.7-linux-amd64.tar.gz`. ### Compile from source code diff --git a/production-deployment-using-tiup.md b/production-deployment-using-tiup.md index 579c3fdf0b939..ee183dd70f59e 100644 --- a/production-deployment-using-tiup.md +++ b/production-deployment-using-tiup.md @@ -333,13 +333,13 @@ Before you run the `deploy` command, use the `check` and `check --apply` command {{< copyable "shell-regular" >}} ```shell - tiup cluster deploy tidb-test v6.5.6 ./topology.yaml --user root [-p] [-i /home/root/.ssh/gcp_rsa] + tiup cluster deploy tidb-test v6.5.7 ./topology.yaml --user root [-p] [-i /home/root/.ssh/gcp_rsa] ``` In the `tiup cluster deploy` command above: - `tidb-test` is the name of the TiDB cluster to be deployed. -- `v6.5.6` is the version of the TiDB cluster to be deployed. You can see the latest supported versions by running `tiup list tidb`. +- `v6.5.7` is the version of the TiDB cluster to be deployed. You can see the latest supported versions by running `tiup list tidb`. - `topology.yaml` is the initialization configuration file. - `--user root` indicates logging into the target machine as the `root` user to complete the cluster deployment. The `root` user is expected to have `ssh` and `sudo` privileges to the target machine. Alternatively, you can use other users with `ssh` and `sudo` privileges to complete the deployment. - `[-i]` and `[-p]` are optional. If you have configured login to the target machine without password, these parameters are not required. If not, choose one of the two parameters. `[-i]` is the private key of the root user (or other users specified by `--user`) that has access to the target machine. `[-p]` is used to input the user password interactively. diff --git a/quick-start-with-tidb.md b/quick-start-with-tidb.md index a523501b0f2da..5c4bf78d0d25e 100644 --- a/quick-start-with-tidb.md +++ b/quick-start-with-tidb.md @@ -80,10 +80,10 @@ As a distributed system, a basic TiDB test cluster usually consists of 2 TiDB in {{< copyable "shell-regular" >}} ```shell - tiup playground v6.5.6 --db 2 --pd 3 --kv 3 + tiup playground v6.5.7 --db 2 --pd 3 --kv 3 ``` - The command downloads a version cluster to the local machine and starts it, such as v6.5.6. To view the latest version, run `tiup list tidb`. + The command downloads a version cluster to the local machine and starts it, such as v6.5.7. To view the latest version, run `tiup list tidb`. This command returns the access methods of the cluster: @@ -201,10 +201,10 @@ As a distributed system, a basic TiDB test cluster usually consists of 2 TiDB in {{< copyable "shell-regular" >}} ```shell - tiup playground v6.5.6 --db 2 --pd 3 --kv 3 + tiup playground v6.5.7 --db 2 --pd 3 --kv 3 ``` - The command downloads a version cluster to the local machine and starts it, such as v6.5.6. To view the latest version, run `tiup list tidb`. + The command downloads a version cluster to the local machine and starts it, such as v6.5.7. To view the latest version, run `tiup list tidb`. This command returns the access methods of the cluster: @@ -436,7 +436,7 @@ Other requirements for the target machine: ``` - `<cluster-name>`: Set the cluster name - - `<version>`: Set the TiDB cluster version, such as `v6.5.6`. You can see all the supported TiDB versions by running the `tiup list tidb` command + - `<version>`: Set the TiDB cluster version, such as `v6.5.7`. You can see all the supported TiDB versions by running the `tiup list tidb` command - `-p`: Specify the password used to connect to the target machine. > **Note:** diff --git a/scale-tidb-using-tiup.md b/scale-tidb-using-tiup.md index bb84689705cfe..bb4cb1796b0af 100644 --- a/scale-tidb-using-tiup.md +++ b/scale-tidb-using-tiup.md @@ -275,7 +275,7 @@ This section exemplifies how to remove a TiKV node from the `10.0.1.5` host. ``` Starting /root/.tiup/components/cluster/v1.11.3/cluster display <cluster-name> TiDB Cluster: <cluster-name> - TiDB Version: v6.5.6 + TiDB Version: v6.5.7 ID Role Host Ports Status Data Dir Deploy Dir -- ---- ---- ----- ------ -------- ---------- 10.0.1.3:8300 cdc 10.0.1.3 8300 Up data/cdc-8300 deploy/cdc-8300 diff --git a/ticdc/deploy-ticdc.md b/ticdc/deploy-ticdc.md index b96190269990b..16796faa02e1c 100644 --- a/ticdc/deploy-ticdc.md +++ b/ticdc/deploy-ticdc.md @@ -95,7 +95,7 @@ tiup cluster upgrade <cluster-name> <version> --transfer-timeout 600 > **Note:** > -> In the preceding command, you need to replace `<cluster-name>` and `<version>` with the actual cluster name and cluster version. For example, the version can be `v6.5.6`. +> In the preceding command, you need to replace `<cluster-name>` and `<version>` with the actual cluster name and cluster version. For example, the version can be `v6.5.7`. ### Upgrade cautions @@ -152,7 +152,7 @@ See [Enable TLS Between TiDB Components](/enable-tls-between-components.md). ## View TiCDC status using the command-line tool -Run the following command to view the TiCDC cluster status. Note that you need to replace `v<CLUSTER_VERSION>` with the TiCDC cluster version, such as `v6.5.6`: +Run the following command to view the TiCDC cluster status. Note that you need to replace `v<CLUSTER_VERSION>` with the TiCDC cluster version, such as `v6.5.7`: ```shell tiup ctl:v<CLUSTER_VERSION> cdc capture list --server=http://10.0.10.25:8300 diff --git a/ticdc/ticdc-changefeed-config.md b/ticdc/ticdc-changefeed-config.md index 3b458bd435fb4..1e9c069f37c27 100644 --- a/ticdc/ticdc-changefeed-config.md +++ b/ticdc/ticdc-changefeed-config.md @@ -16,7 +16,7 @@ cdc cli changefeed create --server=http://10.0.10.25:8300 --sink-uri="mysql://ro ```shell Create changefeed successfully! ID: simple-replication-task -Info: {"upstream_id":7178706266519722477,"namespace":"default","id":"simple-replication-task","sink_uri":"mysql://root:xxxxx@127.0.0.1:4000/?time-zone=","create_time":"2023-12-07T15:05:46.679218+08:00","start_ts":438156275634929669,"engine":"unified","config":{"case_sensitive":false,"enable_old_value":true,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":true,"bdr_mode":false,"sync_point_interval":30000000000,"sync_point_retention":3600000000000,"filter":{"rules":["test.*"],"event_filters":null},"mounter":{"worker_num":16},"sink":{"protocol":"","schema_registry":"","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false},"column_selectors":null,"transaction_atomicity":"none","encoder_concurrency":16,"terminator":"\r\n","date_separator":"none","enable_partition_separator":false},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"storage":""}},"state":"normal","creator_version":"v6.5.6"} +Info: {"upstream_id":7178706266519722477,"namespace":"default","id":"simple-replication-task","sink_uri":"mysql://root:xxxxx@127.0.0.1:4000/?time-zone=","create_time":"2023-12-07T15:05:46.679218+08:00","start_ts":438156275634929669,"engine":"unified","config":{"case_sensitive":false,"enable_old_value":true,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":true,"bdr_mode":false,"sync_point_interval":30000000000,"sync_point_retention":3600000000000,"filter":{"rules":["test.*"],"event_filters":null},"mounter":{"worker_num":16},"sink":{"protocol":"","schema_registry":"","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false},"column_selectors":null,"transaction_atomicity":"none","encoder_concurrency":16,"terminator":"\r\n","date_separator":"none","enable_partition_separator":false},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"storage":""}},"state":"normal","creator_version":"v6.5.7"} ``` - `--changefeed-id`: The ID of the replication task. The format must match the `^[a-zA-Z0-9]+(\-[a-zA-Z0-9]+)*$` regular expression. If this ID is not specified, TiCDC automatically generates a UUID (the version 4 format) as the ID. @@ -205,7 +205,7 @@ sasl-oauth-grant-type = "client_credentials" sasl-oauth-audience = "kafka" [sink.cloud-storage-config] -# The concurrency for saving data changes to the downstream cloud storage. +# The concurrency for saving data changes to the downstream cloud storage. # The default value is 16. worker-count = 16 # The interval for saving data changes to the downstream cloud storage. @@ -215,7 +215,7 @@ flush-interval = "2s" # The default value is 67108864 (this is, 64 MiB). file-size = 67108864 # The duration to retain files, which takes effect only when `date-separator` is configured as `day`. Assume that `file-expiration-days = 1` and `file-cleanup-cron-spec = "0 0 0 * * *"`, then TiCDC performs daily cleanup at 00:00:00 for files saved beyond 24 hours. For example, at 00:00:00 on 2023/12/02, TiCDC cleans up files generated before 2023/12/01, while files generated on 2023/12/01 remain unaffected. -# The default value is 0, which means file cleanup is disabled. +# The default value is 0, which means file cleanup is disabled. file-expiration-days = 0 # The running cycle of the scheduled cleanup task, compatible with the crontab configuration, with a format of `<Second> <Minute> <Hour> <Day of the month> <Month> <Day of the week (Optional)>` # The default value is "0 0 2 * * *", which means that the cleanup task is executed every day at 2 AM. diff --git a/ticdc/ticdc-changefeed-overview.md b/ticdc/ticdc-changefeed-overview.md index b334ed1033b46..b2756b2e2fbc2 100644 --- a/ticdc/ticdc-changefeed-overview.md +++ b/ticdc/ticdc-changefeed-overview.md @@ -43,4 +43,4 @@ You can manage a TiCDC cluster and its replication tasks using the command-line You can also use the HTTP interface (the TiCDC OpenAPI feature) to manage a TiCDC cluster and its replication tasks. For details, see [TiCDC OpenAPI](/ticdc/ticdc-open-api.md). -If your TiCDC is deployed using TiUP, you can start `cdc cli` by running the `tiup ctl:v<CLUSTER_VERSION> cdc` command. Replace `v<CLUSTER_VERSION>` with the TiCDC cluster version, such as `v6.5.6`. You can also run `cdc cli` directly. +If your TiCDC is deployed using TiUP, you can start `cdc cli` by running the `tiup ctl:v<CLUSTER_VERSION> cdc` command. Replace `v<CLUSTER_VERSION>` with the TiCDC cluster version, such as `v6.5.7`. You can also run `cdc cli` directly. diff --git a/ticdc/ticdc-manage-changefeed.md b/ticdc/ticdc-manage-changefeed.md index 5608f32f96570..66079922e613d 100644 --- a/ticdc/ticdc-manage-changefeed.md +++ b/ticdc/ticdc-manage-changefeed.md @@ -19,7 +19,7 @@ cdc cli changefeed create --server=http://10.0.10.25:8300 --sink-uri="mysql://ro ```shell Create changefeed successfully! ID: simple-replication-task -Info: {"upstream_id":7178706266519722477,"namespace":"default","id":"simple-replication-task","sink_uri":"mysql://root:xxxxx@127.0.0.1:4000/?time-zone=","create_time":"2023-12-07T15:05:46.679218+08:00","start_ts":438156275634929669,"engine":"unified","config":{"case_sensitive":false,"enable_old_value":true,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":true,"bdr_mode":false,"sync_point_interval":30000000000,"sync_point_retention":3600000000000,"filter":{"rules":["test.*"],"event_filters":null},"mounter":{"worker_num":16},"sink":{"protocol":"","schema_registry":"","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false},"column_selectors":null,"transaction_atomicity":"none","encoder_concurrency":16,"terminator":"\r\n","date_separator":"none","enable_partition_separator":false},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"storage":""}},"state":"normal","creator_version":"v6.5.6"} +Info: {"upstream_id":7178706266519722477,"namespace":"default","id":"simple-replication-task","sink_uri":"mysql://root:xxxxx@127.0.0.1:4000/?time-zone=","create_time":"2023-12-07T15:05:46.679218+08:00","start_ts":438156275634929669,"engine":"unified","config":{"case_sensitive":false,"enable_old_value":true,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":true,"bdr_mode":false,"sync_point_interval":30000000000,"sync_point_retention":3600000000000,"filter":{"rules":["test.*"],"event_filters":null},"mounter":{"worker_num":16},"sink":{"protocol":"","schema_registry":"","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false},"column_selectors":null,"transaction_atomicity":"none","encoder_concurrency":16,"terminator":"\r\n","date_separator":"none","enable_partition_separator":false},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"storage":""}},"state":"normal","creator_version":"v6.5.7"} ``` ## Query the replication task list diff --git a/ticdc/ticdc-sink-to-cloud-storage.md b/ticdc/ticdc-sink-to-cloud-storage.md index cae1a91968f2b..b213c93945cc5 100644 --- a/ticdc/ticdc-sink-to-cloud-storage.md +++ b/ticdc/ticdc-sink-to-cloud-storage.md @@ -28,7 +28,7 @@ cdc cli changefeed create \ The output is as follows: ```shell -Info: {"upstream_id":7171388873935111376,"namespace":"default","id":"simple-replication-task","sink_uri":"s3://logbucket/storage_test?protocol=canal-json","create_time":"2023-12-07T18:52:05.566016967+08:00","start_ts":437706850431664129,"engine":"unified","config":{"case_sensitive":false,"enable_old_value":true,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"],"event_filters":null},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","schema_registry":"","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false},"column_selectors":null,"transaction_atomicity":"none","encoder_concurrency":16,"terminator":"\r\n","date_separator":"none","enable_partition_separator":false},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"storage":""}},"state":"normal","creator_version":"v6.5.6"} +Info: {"upstream_id":7171388873935111376,"namespace":"default","id":"simple-replication-task","sink_uri":"s3://logbucket/storage_test?protocol=canal-json","create_time":"2023-12-07T18:52:05.566016967+08:00","start_ts":437706850431664129,"engine":"unified","config":{"case_sensitive":false,"enable_old_value":true,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"],"event_filters":null},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","schema_registry":"","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false},"column_selectors":null,"transaction_atomicity":"none","encoder_concurrency":16,"terminator":"\r\n","date_separator":"none","enable_partition_separator":false},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"storage":""}},"state":"normal","creator_version":"v6.5.7"} ``` - `--server`: The address of any TiCDC server in the TiCDC cluster. diff --git a/tidb-binlog/get-started-with-tidb-binlog.md b/tidb-binlog/get-started-with-tidb-binlog.md index b21e52073c9bf..074d71ee6a593 100644 --- a/tidb-binlog/get-started-with-tidb-binlog.md +++ b/tidb-binlog/get-started-with-tidb-binlog.md @@ -42,7 +42,7 @@ sudo yum install -y mariadb-server ``` ```bash -curl -L https://download.pingcap.org/tidb-community-server-v6.5.6-linux-amd64.tar.gz | tar xzf - +curl -L https://download.pingcap.org/tidb-community-server-v6.5.7-linux-amd64.tar.gz | tar xzf - cd tidb-latest-linux-amd64 ``` diff --git a/tiflash/create-tiflash-replicas.md b/tiflash/create-tiflash-replicas.md index 81afe54bd4c12..8a308aa45e2c6 100644 --- a/tiflash/create-tiflash-replicas.md +++ b/tiflash/create-tiflash-replicas.md @@ -143,7 +143,7 @@ Before TiFlash replicas are added, each TiKV instance performs a full table scan tiup ctl:v<CLUSTER_VERSION> pd -u http://<PD_ADDRESS>:2379 store limit all engine tiflash 60 add-peer ``` - > In the preceding command, you need to replace `v<CLUSTER_VERSION>` with the actual cluster version, such as `v6.5.6` and `<PD_ADDRESS>:2379` with the address of any PD node. For example: + > In the preceding command, you need to replace `v<CLUSTER_VERSION>` with the actual cluster version, such as `v6.5.7` and `<PD_ADDRESS>:2379` with the address of any PD node. For example: > > ```shell > tiup ctl:v6.1.1 pd -u http://192.168.1.4:2379 store limit all engine tiflash 60 add-peer diff --git a/tiup/tiup-cluster.md b/tiup/tiup-cluster.md index 3974d2e721b59..0d4eb6c0677f0 100644 --- a/tiup/tiup-cluster.md +++ b/tiup/tiup-cluster.md @@ -61,7 +61,7 @@ To deploy the cluster, run the `tiup cluster deploy` command. The usage of the c tiup cluster deploy <cluster-name> <version> <topology.yaml> [flags] ``` -This command requires you to provide the cluster name, the TiDB cluster version (such as `v6.5.6`), and a topology file of the cluster. +This command requires you to provide the cluster name, the TiDB cluster version (such as `v6.5.7`), and a topology file of the cluster. To write a topology file, refer to [the example](https://github.com/pingcap/tiup/blob/master/embed/examples/cluster/topology.example.yaml). The following file is an example of the simplest topology: @@ -118,12 +118,12 @@ tidb_servers: ... ``` -Save the file as `/tmp/topology.yaml`. If you want to use TiDB v6.5.6 and your cluster name is `prod-cluster`, run the following command: +Save the file as `/tmp/topology.yaml`. If you want to use TiDB v6.5.7 and your cluster name is `prod-cluster`, run the following command: {{< copyable "shell-regular" >}} ```shell -tiup cluster deploy -p prod-cluster v6.5.6 /tmp/topology.yaml +tiup cluster deploy -p prod-cluster v6.5.7 /tmp/topology.yaml ``` During the execution, TiUP asks you to confirm your topology again and requires the root password of the target machine (the `-p` flag means inputting password): @@ -131,7 +131,7 @@ During the execution, TiUP asks you to confirm your topology again and requires ```bash Please confirm your topology: TiDB Cluster: prod-cluster -TiDB Version: v6.5.6 +TiDB Version: v6.5.7 Type Host Ports OS/Arch Directories ---- ---- ----- ------- ----------- pd 172.16.5.134 2379/2380 linux/x86_64 deploy/pd-2379,data/pd-2379 @@ -174,7 +174,7 @@ tiup cluster list Starting /root/.tiup/components/cluster/v1.11.3/cluster list Name User Version Path PrivateKey ---- ---- ------- ---- ---------- -prod-cluster tidb v6.5.6 /root/.tiup/storage/cluster/clusters/prod-cluster /root/.tiup/storage/cluster/clusters/prod-cluster/ssh/id_rsa +prod-cluster tidb v6.5.7 /root/.tiup/storage/cluster/clusters/prod-cluster /root/.tiup/storage/cluster/clusters/prod-cluster/ssh/id_rsa ``` ## Start the cluster @@ -204,7 +204,7 @@ tiup cluster display prod-cluster ``` Starting /root/.tiup/components/cluster/v1.11.3/cluster display prod-cluster TiDB Cluster: prod-cluster -TiDB Version: v6.5.6 +TiDB Version: v6.5.7 ID Role Host Ports OS/Arch Status Data Dir Deploy Dir -- ---- ---- ----- ------- ------ -------- ---------- 172.16.5.134:3000 grafana 172.16.5.134 3000 linux/x86_64 Up - deploy/grafana-3000 @@ -278,7 +278,7 @@ tiup cluster display prod-cluster ``` Starting /root/.tiup/components/cluster/v1.11.3/cluster display prod-cluster TiDB Cluster: prod-cluster -TiDB Version: v6.5.6 +TiDB Version: v6.5.7 ID Role Host Ports OS/Arch Status Data Dir Deploy Dir -- ---- ---- ----- ------- ------ -------- ---------- 172.16.5.134:3000 grafana 172.16.5.134 3000 linux/x86_64 Up - deploy/grafana-3000 @@ -389,12 +389,12 @@ Global Flags: -y, --yes Skip all confirmations and assumes 'yes' ``` -For example, the following command upgrades the cluster to v6.5.6: +For example, the following command upgrades the cluster to v6.5.7: {{< copyable "shell-regular" >}} ```bash -tiup cluster upgrade tidb-test v6.5.6 +tiup cluster upgrade tidb-test v6.5.7 ``` ## Update configuration @@ -579,11 +579,11 @@ tiup cluster audit Starting component `cluster`: /home/tidb/.tiup/components/cluster/v1.11.3/cluster audit ID Time Command -- ---- ------- -4BLhr0 2023-12-07T13:25:09+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.6 /tmp/topology.yaml -4BKWjF 2023-12-07T23:36:57+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.6 /tmp/topology.yaml -4BKVwH 2023-12-07T23:02:08+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.6 /tmp/topology.yaml -4BKKH1 2023-12-07T16:39:04+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster destroy test -4BKKDx 2023-12-07T16:36:57+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.6 /tmp/topology.yaml +4BLhr0 2024-01-08T13:25:09+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.7 /tmp/topology.yaml +4BKWjF 2024-01-08T23:36:57+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.7 /tmp/topology.yaml +4BKVwH 2024-01-08T23:02:08+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.7 /tmp/topology.yaml +4BKKH1 2024-01-08T16:39:04+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster destroy test +4BKKDx 2024-01-08T16:36:57+08:00 /home/tidb/.tiup/components/cluster/v1.11.3/cluster deploy test v6.5.7 /tmp/topology.yaml ``` The first column is `audit-id`. To view the execution log of a certain command, pass the `audit-id` of a command as the flag as follows: @@ -699,7 +699,7 @@ All operations above performed on the cluster machine use the SSH client embedde Then you can use the `--ssh=system` command-line flag to enable the system-native command-line tool: -- Deploy a cluster: `tiup cluster deploy <cluster-name> <version> <topo> --ssh=system`. Fill in the name of your cluster for `<cluster-name>`, the TiDB version to be deployed (such as `v6.5.6`) for `<version>`, and the topology file for `<topo>`. +- Deploy a cluster: `tiup cluster deploy <cluster-name> <version> <topo> --ssh=system`. Fill in the name of your cluster for `<cluster-name>`, the TiDB version to be deployed (such as `v6.5.7`) for `<version>`, and the topology file for `<topo>`. - Start a cluster: `tiup cluster start <cluster-name> --ssh=system` - Upgrade a cluster: `tiup cluster upgrade ... --ssh=system` diff --git a/tiup/tiup-component-cluster-deploy.md b/tiup/tiup-component-cluster-deploy.md index a460532618a0a..004d5e11bb955 100644 --- a/tiup/tiup-component-cluster-deploy.md +++ b/tiup/tiup-component-cluster-deploy.md @@ -13,7 +13,7 @@ tiup cluster deploy <cluster-name> <version> <topology.yaml> [flags] ``` - `<cluster-name>`: the name of the new cluster, which cannot be the same as the existing cluster names. -- `<version>`: the version number of the TiDB cluster to deploy, such as `v6.5.6`. +- `<version>`: the version number of the TiDB cluster to deploy, such as `v6.5.7`. - `<topology.yaml>`: the prepared [topology file](/tiup/tiup-cluster-topology-reference.md). ## Options diff --git a/tiup/tiup-component-cluster-patch.md b/tiup/tiup-component-cluster-patch.md index 636e420ba7976..3c6f5bb4404b8 100644 --- a/tiup/tiup-component-cluster-patch.md +++ b/tiup/tiup-component-cluster-patch.md @@ -28,7 +28,7 @@ Before running the `tiup cluster patch` command, you need to pack the binary pac 1. Determine the following variables: - `${component}`: the name of the component to be replaced (such as `tidb`, `tikv`, or `pd`). - - `${version}`: the version of the component (such as `v6.5.6`). + - `${version}`: the version of the component (such as `v6.5.7`). - `${os}`: the operating system (`linux`). - `${arch}`: the platform on which the component runs (`amd64`, `arm64`). diff --git a/tiup/tiup-component-cluster-upgrade.md b/tiup/tiup-component-cluster-upgrade.md index 457a08128fd3d..889323109d975 100644 --- a/tiup/tiup-component-cluster-upgrade.md +++ b/tiup/tiup-component-cluster-upgrade.md @@ -13,7 +13,7 @@ tiup cluster upgrade <cluster-name> <version> [flags] ``` - `<cluster-name>`: the cluster name to operate on. If you forget the cluster name, you can check it with the [cluster list](/tiup/tiup-component-cluster-list.md) command. -- `<version>`: the target version to upgrade to, such as `v6.5.6`. Currently, it is only allowed to upgrade to a version higher than the current cluster, that is, no downgrade is allowed. It is also not allowed to upgrade to the nightly version. +- `<version>`: the target version to upgrade to, such as `v6.5.7`. Currently, it is only allowed to upgrade to a version higher than the current cluster, that is, no downgrade is allowed. It is also not allowed to upgrade to the nightly version. ## Options diff --git a/tiup/tiup-component-dm-upgrade.md b/tiup/tiup-component-dm-upgrade.md index 2062c41db52f9..3a3b54ce35d33 100644 --- a/tiup/tiup-component-dm-upgrade.md +++ b/tiup/tiup-component-dm-upgrade.md @@ -13,7 +13,7 @@ tiup dm upgrade <cluster-name> <version> [flags] ``` - `<cluster-name>` is the name of the cluster to be operated on. If you forget the cluster name, you can check it using the [`tiup dm list`](/tiup/tiup-component-dm-list.md) command. -- `<version>` is the target version to be upgraded to, such as `v6.5.6`. Currently, only upgrading to a later version is allowed, and upgrading to an earlier version is not allowed, which means the downgrade is not allowed. Upgrading to a nightly version is not allowed either. +- `<version>` is the target version to be upgraded to, such as `v6.5.7`. Currently, only upgrading to a later version is allowed, and upgrading to an earlier version is not allowed, which means the downgrade is not allowed. Upgrading to a nightly version is not allowed either. ## Options diff --git a/tiup/tiup-component-management.md b/tiup/tiup-component-management.md index 5ba4a018cbae9..02cc715f75ec7 100644 --- a/tiup/tiup-component-management.md +++ b/tiup/tiup-component-management.md @@ -69,12 +69,12 @@ Example 2: Use TiUP to install the nightly version of TiDB. tiup install tidb:nightly ``` -Example 3: Use TiUP to install TiKV v6.5.6. +Example 3: Use TiUP to install TiKV v6.5.7. {{< copyable "shell-regular" >}} ```shell -tiup install tikv:v6.5.6 +tiup install tikv:v6.5.7 ``` ## Upgrade components @@ -127,12 +127,12 @@ Before the component is started, TiUP creates a directory for it, and then puts If you want to start the same component multiple times and reuse the previous working directory, you can use `--tag` to specify the same name when the component is started. After the tag is specified, the working directory will *not be automatically deleted* when the instance is terminated, which makes it convenient to reuse the working directory. -Example 1: Operate TiDB v6.5.6. +Example 1: Operate TiDB v6.5.7. {{< copyable "shell-regular" >}} ```shell -tiup tidb:v6.5.6 +tiup tidb:v6.5.7 ``` Example 2: Specify the tag with which TiKV operates. @@ -218,12 +218,12 @@ The following flags are supported in this command: - If the version is ignored, adding `--all` means to uninstall all versions of this component. - If the version and the component are both ignored, adding `--all` means to uninstall all components of all versions. -Example 1: Uninstall TiDB v6.5.6. +Example 1: Uninstall TiDB v6.5.7. {{< copyable "shell-regular" >}} ```shell -tiup uninstall tidb:v6.5.6 +tiup uninstall tidb:v6.5.7 ``` Example 2: Uninstall TiKV of all versions. diff --git a/tiup/tiup-mirror.md b/tiup/tiup-mirror.md index 617fea63e08d1..336ce59ec8b77 100644 --- a/tiup/tiup-mirror.md +++ b/tiup/tiup-mirror.md @@ -86,9 +86,9 @@ The `tiup mirror clone` command provides many optional flags (might provide more If you want to clone only one version (not all versions) of a component, use `--<component>=<version>` to specify this version. For example: - - Execute the `tiup mirror clone <target-dir> --tidb v6.5.6` command to clone the v6.5.6 version of the TiDB component. - - Run the `tiup mirror clone <target-dir> --tidb v6.5.6 --tikv all` command to clone the v6.5.6 version of the TiDB component and all versions of the TiKV component. - - Run the `tiup mirror clone <target-dir> v6.5.6` command to clone the v6.5.6 version of all components in a cluster. + - Execute the `tiup mirror clone <target-dir> --tidb v6.5.7` command to clone the v6.5.7 version of the TiDB component. + - Run the `tiup mirror clone <target-dir> --tidb v6.5.7 --tikv all` command to clone the v6.5.7 version of the TiDB component and all versions of the TiKV component. + - Run the `tiup mirror clone <target-dir> v6.5.7` command to clone the v6.5.7 version of all components in a cluster. After cloning, signing keys are set up automatically. diff --git a/upgrade-tidb-using-tiup.md b/upgrade-tidb-using-tiup.md index 00373d63a827c..64b3af2f4bfb4 100644 --- a/upgrade-tidb-using-tiup.md +++ b/upgrade-tidb-using-tiup.md @@ -31,12 +31,12 @@ This document is targeted for the following upgrade paths: ## Upgrade caveat - TiDB currently does not support version downgrade or rolling back to an earlier version after the upgrade. -- For the v4.0 cluster managed using TiDB Ansible, you need to import the cluster to TiUP (`tiup cluster`) for new management according to [Upgrade TiDB Using TiUP (v4.0)](https://docs.pingcap.com/tidb/v4.0/upgrade-tidb-using-tiup#import-tidb-ansible-and-the-inventoryini-configuration-to-tiup). Then you can upgrade the cluster to v6.5.6 according to this document. -- To update versions earlier than v3.0 to v6.5.6: +- For the v4.0 cluster managed using TiDB Ansible, you need to import the cluster to TiUP (`tiup cluster`) for new management according to [Upgrade TiDB Using TiUP (v4.0)](https://docs.pingcap.com/tidb/v4.0/upgrade-tidb-using-tiup#import-tidb-ansible-and-the-inventoryini-configuration-to-tiup). Then you can upgrade the cluster to v6.5.7 according to this document. +- To update versions earlier than v3.0 to v6.5.7: 1. Update this version to 3.0 using [TiDB Ansible](https://docs.pingcap.com/tidb/v3.0/upgrade-tidb-using-ansible). 2. Use TiUP (`tiup cluster`) to import the TiDB Ansible configuration. 3. Update the 3.0 version to 4.0 according to [Upgrade TiDB Using TiUP (v4.0)](https://docs.pingcap.com/tidb/v4.0/upgrade-tidb-using-tiup#import-tidb-ansible-and-the-inventoryini-configuration-to-tiup). - 4. Upgrade the cluster to v6.5.6 according to this document. + 4. Upgrade the cluster to v6.5.7 according to this document. - Support upgrading the versions of TiDB Binlog, TiCDC, TiFlash, and other components. - When upgrading TiFlash from versions earlier than v6.3.0 to v6.3.0 and later versions, note that the CPU must support the AVX2 instruction set under the Linux AMD64 architecture and the ARMv8 instruction set architecture under the Linux ARM64 architecture. For details, see the description in [v6.3.0 Release Notes](/releases/release-6.3.0.md#others). - For detailed compatibility changes of different versions, see the [Release Notes](/releases/release-notes.md) of each version. Modify your cluster configuration according to the "Compatibility Changes" section of the corresponding release notes. @@ -50,7 +50,7 @@ This section introduces the preparation works needed before upgrading your TiDB Review compatibility changes in TiDB release notes. If any changes affect your upgrade, take actions accordingly. -This following provides compatibility changes you need to know when you upgrade from v6.4.0 to the current version (v6.5.6). If you are upgrading from v6.3.0 or earlier versions to the current version, you might also need to check the compatibility changes introduced in intermediate versions in the corresponding [release notes](/releases/release-notes.md). +This following provides compatibility changes you need to know when you upgrade from v6.4.0 to the current version (v6.5.7). If you are upgrading from v6.3.0 or earlier versions to the current version, you might also need to check the compatibility changes introduced in intermediate versions in the corresponding [release notes](/releases/release-notes.md). - TiDB v6.5.0 [compatibility changes](/releases/release-6.5.0.md#compatibility-changes) and [deprecated features](/releases/release-6.5.0.md#deprecated-feature) - TiDB v6.5.1 [compatibility changes](/releases/release-6.5.1.md#compatibility-changes) @@ -59,6 +59,7 @@ This following provides compatibility changes you need to know when you upgrade - TiDB v6.5.4 [release notes](/releases/release-6.5.4.md) - TiDB v6.5.5 [release notes](/releases/release-6.5.5.md) - TiDB v6.5.6 [release notes](/releases/release-6.5.6.md) +- TiDB v6.5.7 [release notes](/releases/release-6.5.7.md) ### Step 2: Upgrade TiUP or TiUP offline mirror @@ -133,7 +134,7 @@ Now, the offline mirror has been upgraded successfully. If an error occurs durin > Skip this step if one of the following situations applies: > > + You have not modified the configuration parameters of the original cluster. Or you have modified the configuration parameters using `tiup cluster` but no more modification is needed. -> + After the upgrade, you want to use v6.5.6's default parameter values for the unmodified configuration items. +> + After the upgrade, you want to use v6.5.7's default parameter values for the unmodified configuration items. 1. Enter the `vi` editing mode to edit the topology file: @@ -149,7 +150,7 @@ Now, the offline mirror has been upgraded successfully. If an error occurs durin > **Note:** > -> Before you upgrade the cluster to v6.5.6, make sure that the parameters you have modified in v4.0 are compatible in v6.5.6. For details, see [TiKV Configuration File](/tikv-configuration-file.md). +> Before you upgrade the cluster to v6.5.7, make sure that the parameters you have modified in v4.0 are compatible in v6.5.7. For details, see [TiKV Configuration File](/tikv-configuration-file.md). ### Step 4: Check the health status of the current cluster @@ -193,12 +194,12 @@ If your application has a maintenance window for the database to be stopped for tiup cluster upgrade <cluster-name> <version> ``` -For example, if you want to upgrade the cluster to v6.5.6: +For example, if you want to upgrade the cluster to v6.5.7: {{< copyable "shell-regular" >}} ```shell -tiup cluster upgrade <cluster-name> v6.5.6 +tiup cluster upgrade <cluster-name> v6.5.7 ``` > **Note:** @@ -226,7 +227,7 @@ tiup cluster upgrade <cluster-name> v6.5.6 tiup cluster stop <cluster-name> ``` -2. Use the `upgrade` command with the `--offline` option to perform the offline upgrade. Fill in the name of your cluster for `<cluster-name>` and the version to upgrade to for `<version>`, such as `v6.5.6`. +2. Use the `upgrade` command with the `--offline` option to perform the offline upgrade. Fill in the name of your cluster for `<cluster-name>` and the version to upgrade to for `<version>`, such as `v6.5.7`. {{< copyable "shell-regular" >}} @@ -255,7 +256,7 @@ tiup cluster display <cluster-name> ``` Cluster type: tidb Cluster name: <cluster-name> -Cluster version: v6.5.6 +Cluster version: v6.5.7 ``` ## FAQ @@ -310,7 +311,7 @@ Starting from v6.2.0, TiDB enables the [concurrent DDL framework](/ddl-introduct ### The evict leader has waited too long during the upgrade. How to skip this step for a quick upgrade? -You can specify `--force`. Then the processes of transferring PD leader and evicting TiKV leader are skipped during the upgrade. The cluster is directly restarted to update the version, which has a great impact on the cluster that runs online. In the following command, `<version>` is the version to upgrade to, such as `v6.5.6`. +You can specify `--force`. Then the processes of transferring PD leader and evicting TiKV leader are skipped during the upgrade. The cluster is directly restarted to update the version, which has a great impact on the cluster that runs online. In the following command, `<version>` is the version to upgrade to, such as `v6.5.7`. {{< copyable "shell-regular" >}} @@ -325,5 +326,5 @@ You can upgrade the tool version by using TiUP to install the `ctl` component of {{< copyable "shell-regular" >}} ```shell -tiup install ctl:v6.5.6 +tiup install ctl:v6.5.7 ``` From c317af332fe404730cc686bdde2c28a2d2fdd512 Mon Sep 17 00:00:00 2001 From: Ran <huangran.alex@gmail.com> Date: Fri, 5 Jan 2024 17:50:06 +0800 Subject: [PATCH 4/4] address comments --- tiup/tiup-playground.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tiup/tiup-playground.md b/tiup/tiup-playground.md index 043c7a855ebbd..c699ce3acab42 100644 --- a/tiup/tiup-playground.md +++ b/tiup/tiup-playground.md @@ -21,7 +21,7 @@ This command actually performs the following operations: - Because this command does not specify the version of the playground component, TiUP first checks the latest version of the installed playground component. Assume that the latest version is v1.11.3, then this command works the same as `tiup playground:v1.11.3`. - If you have not used TiUP playground to install the TiDB, TiKV, and PD components, the playground component installs the latest stable version of these components, and then start these instances. -- Because this command does not specify the version of the TiDB, PD, and TiKV component, TiUP playground uses the latest version of each component by default. Assume that the latest version is v6.5.6, then this command works the same as `tiup playground:v1.11.3 v6.5.6`. +- Because this command does not specify the version of the TiDB, PD, and TiKV component, TiUP playground uses the latest version of each component by default. Assume that the latest version is v6.5.7, then this command works the same as `tiup playground:v1.11.3 v6.5.7`. - Because this command does not specify the number of each component, TiUP playground, by default, starts a smallest cluster that consists of one TiDB instance, one TiKV instance, one PD instance, and one TiFlash instance. - After starting each TiDB component, TiUP playground reminds you that the cluster is successfully started and provides you some useful information, such as how to connect to the TiDB cluster through the MySQL client and how to access the [TiDB Dashboard](/dashboard/dashboard-intro.md).