From 2c3fccd7f1b6e6e55eb32d86f2cf3688087deb3e Mon Sep 17 00:00:00 2001 From: Dustin Brown Date: Wed, 22 Apr 2020 13:18:00 -0700 Subject: [PATCH] benchmark/sqllogictest_tester/run_tester.sh: Cleanup --- .../SQLLogicTestTesterJenkinsfile | 11 +--- benchmark/sqllogictest_tester/run_tester.sh | 58 ++++++++----------- 2 files changed, 28 insertions(+), 41 deletions(-) diff --git a/benchmark/sqllogictest_tester/SQLLogicTestTesterJenkinsfile b/benchmark/sqllogictest_tester/SQLLogicTestTesterJenkinsfile index 4c8713ee32..7ba24ae06c 100644 --- a/benchmark/sqllogictest_tester/SQLLogicTestTesterJenkinsfile +++ b/benchmark/sqllogictest_tester/SQLLogicTestTesterJenkinsfile @@ -44,14 +44,9 @@ pipeline { } } failure { - script { - def watchers = "$TEST_WATCHERS".split() - for (int i = 0; i < watchers.size(); ++i) { - emailext body: "${currentBuild.currentResult}: Job ${env.JOB_NAME} build ${env.BUILD_NUMBER}\n More info at: ${env.BUILD_URL}", - to: watchers[i], - subject: "Jenkins Build ${currentBuild.currentResult}: Job ${env.JOB_NAME}" - } - } + emailext body: "${currentBuild.currentResult}: Job ${env.JOB_NAME} build ${env.BUILD_NUMBER}\n More info at: ${env.BUILD_URL}", + to: "$TEST_WATCHERS", + subject: "Jenkins Build ${currentBuild.currentResult}: Job ${env.JOB_NAME}" } } } diff --git a/benchmark/sqllogictest_tester/run_tester.sh b/benchmark/sqllogictest_tester/run_tester.sh index ae06093c9d..75d0474374 100755 --- a/benchmark/sqllogictest_tester/run_tester.sh +++ b/benchmark/sqllogictest_tester/run_tester.sh @@ -132,17 +132,17 @@ function import_once() { dolt checkout master } -function import_parsed() { - IFS=', ' read -r -a commit_list <<< "$COMMITS_TO_TEST" - for c in "${commit_list[@]}" - do - seq 1 $TEST_N_TIMES | while read test_num; do - import_once "$c" "$test_num" - done - done -} +#function import_parsed() { +# IFS=', ' read -r -a commit_list <<< "$COMMITS_TO_TEST" +# for c in "${commit_list[@]}" +# do +# seq 1 $TEST_N_TIMES | while read test_num; do +# import_once "$c" "$test_num" +# done +# done +#} -function create_once() { +function create_mean_csv_once() { local commit_hash="$1" exists=$(dolt branch --list "temp-$commit_hash"| sed '/^\s*$/d' | wc -l | tr -d '[:space:]') @@ -160,13 +160,13 @@ function create_once() { dolt checkout master } -function create_committers_mean_csv() { - IFS=', ' read -r -a commit_list <<< "$COMMITS_TO_TEST" - for c in "${commit_list[@]}" - do - create_once "$c" - done -} +#function create_committers_mean_csv() { +# IFS=', ' read -r -a commit_list <<< "$COMMITS_TO_TEST" +# for c in "${commit_list[@]}" +# do +# create_mean_csv_once "$c" +# done +#} function import_and_query_once() { rm -f query_db @@ -197,13 +197,13 @@ SQL if [ "$duration_regressions" != 0 ]; then echo "Duration regression found, $duration_regressions != 0" && echo $duration_query_output && exit 1; else echo "No duration regressions found"; fi } -function import_and_query() { - IFS=', ' read -r -a commit_list <<< "$COMMITS_TO_TEST" - for c in "${commit_list[@]}" - do - import_and_query_once "$c" - done -} +#function import_and_query() { +# IFS=', ' read -r -a commit_list <<< "$COMMITS_TO_TEST" +# for c in "${commit_list[@]}" +# do +# import_and_query_once "$c" +# done +#} function run_once() { local commit_hash="$1" @@ -237,12 +237,10 @@ function run() { run_once "$c" "$test_num" done - (with_dolt_checkout; cd "$dsp_dir"/dolt-sql-performance; create_once "$c") + (with_dolt_checkout; cd "$dsp_dir"/dolt-sql-performance; create_mean_csv_once "$c") (import_and_query_once "$c") done - - rm -rf .dolt } append() { @@ -304,9 +302,3 @@ rm -rf dolt-sql-performance && mkdir dolt-sql-performance (with_dolt_checkout; cd dolt-sql-performance; create_releases_csv) (cd "$logictest_main"; run) - -#(with_dolt_checkout; cd "$dsp_dir"/dolt-sql-performance; import_parsed; create_committers_mean_csv; echo "here are all csvs:"; ls "$TMP_CSV_DIR") - -(echo "here are all csvs:"; ls "$TMP_CSV_DIR") - -#(import_and_query)