github.com/pingcap/ticdc@v0.0.0-20220526033649-485a10ef2652/tests/ddl_sequence/run.sh (about)

     1  #!/bin/bash
     2  
     3  set -e
     4  
     5  CUR=$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )
     6  source $CUR/../_utils/test_prepare
     7  WORK_DIR=$OUT_DIR/$TEST_NAME
     8  CDC_BINARY=cdc.test
     9  SINK_TYPE=$1
    10  
    11  function run() {
    12      rm -rf $WORK_DIR && mkdir -p $WORK_DIR
    13  
    14      start_tidb_cluster --workdir $WORK_DIR
    15  
    16      cd $WORK_DIR
    17  
    18      # record tso before we create tables to skip the system table DDLs
    19      start_ts=$(run_cdc_cli tso query --pd=http://$UP_PD_HOST_1:$UP_PD_PORT_1)
    20  
    21      run_cdc_server --workdir $WORK_DIR --binary $CDC_BINARY
    22  
    23      TOPIC_NAME="ticdc-ddl-sequence-test-$RANDOM"
    24      case $SINK_TYPE in
    25          kafka) SINK_URI="kafka://127.0.0.1:9092/$TOPIC_NAME?partition-num=4&kafka-version=${KAFKA_VERSION}";;
    26          *) SINK_URI="mysql://normal:123456@127.0.0.1:3306/";;
    27      esac
    28      run_cdc_cli changefeed create --start-ts=$start_ts --sink-uri="$SINK_URI"
    29      if [ "$SINK_TYPE" == "kafka" ]; then
    30        run_kafka_consumer $WORK_DIR "kafka://127.0.0.1:9092/$TOPIC_NAME?partition-num=4&version=${KAFKA_VERSION}"
    31      fi
    32      run_sql_file $CUR/data/prepare.sql ${UP_TIDB_HOST} ${UP_TIDB_PORT}
    33      # sync_diff can't check non-exist table, so we check expected tables are created in downstream first
    34      check_table_exists ddl_sequence.finish_mark ${DOWN_TIDB_HOST} ${DOWN_TIDB_PORT}
    35      check_sync_diff $WORK_DIR $CUR/conf/diff_config.toml
    36  
    37      cleanup_process $CDC_BINARY
    38  }
    39  
    40  trap stop_tidb_cluster EXIT
    41  run $*
    42  check_logs $WORK_DIR
    43  echo "[$(date)] <<<<<< run test case $TEST_NAME success! >>>>>>"