From 1b30683e9574965ff3bbff12f69e8eede9c99540 Mon Sep 17 00:00:00 2001 From: Tad Fisher Date: Sat, 12 Feb 2022 19:12:33 -0800 Subject: Update sharness --- test/Makefile | 2 +- test/aggregate-results.sh | 28 +- test/lib-sharness/functions.sh | 593 +++++++++++++++++++++++++++++++ test/sharness.sh | 775 +++++++++++++---------------------------- 4 files changed, 842 insertions(+), 556 deletions(-) create mode 100644 test/lib-sharness/functions.sh (limited to 'test') diff --git a/test/Makefile b/test/Makefile index 16da562..d584e11 100644 --- a/test/Makefile +++ b/test/Makefile @@ -17,7 +17,7 @@ # You should have received a copy of the GNU General Public License # along with this program. If not, see http://www.gnu.org/licenses/ . -SHELL := /bin/bash +SHELL ?= /bin/bash SHELL_PATH ?= $(SHELL) SHELL_PATH_SQ = $(subst ','\'',$(SHELL_PATH)) RM ?= rm -f diff --git a/test/aggregate-results.sh b/test/aggregate-results.sh index 958cbea..cfb766e 100755 --- a/test/aggregate-results.sh +++ b/test/aggregate-results.sh @@ -22,36 +22,36 @@ failed=0 broken=0 total=0 -while read file; do - while read type value; do +while read -r file; do + while read -r type value; do case $type in '') continue ;; fixed) - fixed=$(($fixed + $value)) ;; + fixed=$((fixed + value)) ;; success) - success=$(($success + $value)) ;; + success=$((success + value)) ;; failed) - failed=$(($failed + $value)) - if test $value != 0; then + failed=$((failed + value)) + if test "$value" != 0; then test_name=$(expr "$file" : 'test-results/\(.*\)\.[0-9]*\.counts') failed_tests="$failed_tests $test_name" fi ;; broken) - broken=$(($broken + $value)) ;; + broken=$((broken + value)) ;; total) - total=$(($total + $value)) ;; + total=$((total + value)) ;; esac done <"$file" done if test -n "$failed_tests"; then - printf "\nfailed test(s):$failed_tests\n\n" + printf '\nfailed test(s):%s\n\n' "$failed_tests" fi -printf "%-8s%d\n" fixed $fixed -printf "%-8s%d\n" success $success -printf "%-8s%d\n" failed $failed -printf "%-8s%d\n" broken $broken -printf "%-8s%d\n" total $total +printf '%-8s%d\n' fixed $fixed +printf '%-8s%d\n' success $success +printf '%-8s%d\n' failed $failed +printf '%-8s%d\n' broken $broken +printf '%-8s%d\n' total $total diff --git a/test/lib-sharness/functions.sh b/test/lib-sharness/functions.sh new file mode 100644 index 0000000..56401d8 --- /dev/null +++ b/test/lib-sharness/functions.sh @@ -0,0 +1,593 @@ +# Library of functions shared by all tests scripts, included by +# sharness.sh. +# +# Copyright (c) 2005-2019 Junio C Hamano +# Copyright (c) 2005-2019 Git project +# Copyright (c) 2011-2019 Mathias Lafeldt +# Copyright (c) 2015-2019 Christian Couder +# +# This program is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 2 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see http://www.gnu.org/licenses/ . + +# These assignments are to make shellcheck happy. They should be +# removed when we can use a new version of shellcheck that contains: +# https://github.com/koalaman/shellcheck/pull/1553 +: "${debug:=}" +: "${verbose:=}" +: "${this_test:=}" +: "${skip_all:=}" +: "${EXIT_OK:=}" +: "${test_failure:=0}" +: "${test_fixed:=0}" +: "${test_broken:=0}" +: "${test_success:=0}" + +# Public: Define that a test prerequisite is available. +# +# The prerequisite can later be checked explicitly using test_have_prereq or +# implicitly by specifying the prerequisite name in calls to test_expect_success +# or test_expect_failure. +# +# $1 - Name of prerequisite (a simple word, in all capital letters by convention) +# +# Examples +# +# # Set PYTHON prerequisite if interpreter is available. +# command -v python >/dev/null && test_set_prereq PYTHON +# +# # Set prerequisite depending on some variable. +# test -z "$NO_GETTEXT" && test_set_prereq GETTEXT +# +# Returns nothing. +test_set_prereq() { + satisfied_prereq="$satisfied_prereq$1 " +} +satisfied_prereq=" " + +# Public: Check if one or more test prerequisites are defined. +# +# The prerequisites must have previously been set with test_set_prereq. +# The most common use of this is to skip all the tests if some essential +# prerequisite is missing. +# +# $1 - Comma-separated list of test prerequisites. +# +# Examples +# +# # Skip all remaining tests if prerequisite is not set. +# if ! test_have_prereq PERL; then +# skip_all='skipping perl interface tests, perl not available' +# test_done +# fi +# +# Returns 0 if all prerequisites are defined or 1 otherwise. +test_have_prereq() { + # prerequisites can be concatenated with ',' + save_IFS=$IFS + IFS=, + set -- $@ + IFS=$save_IFS + + total_prereq=0 + ok_prereq=0 + missing_prereq= + + for prerequisite; do + case "$prerequisite" in + !*) + negative_prereq=t + prerequisite=${prerequisite#!} + ;; + *) + negative_prereq= + esac + + total_prereq=$((total_prereq + 1)) + case "$satisfied_prereq" in + *" $prerequisite "*) + satisfied_this_prereq=t + ;; + *) + satisfied_this_prereq= + esac + + case "$satisfied_this_prereq,$negative_prereq" in + t,|,t) + ok_prereq=$((ok_prereq + 1)) + ;; + *) + # Keep a list of missing prerequisites; restore + # the negative marker if necessary. + prerequisite=${negative_prereq:+!}$prerequisite + if test -z "$missing_prereq"; then + missing_prereq=$prerequisite + else + missing_prereq="$prerequisite,$missing_prereq" + fi + esac + done + + test $total_prereq = $ok_prereq +} + +# Public: Execute commands in debug mode. +# +# Takes a single argument and evaluates it only when the test script is started +# with --debug. This is primarily meant for use during the development of test +# scripts. +# +# $1 - Commands to be executed. +# +# Examples +# +# test_debug "cat some_log_file" +# +# Returns the exit code of the last command executed in debug mode or 0 +# otherwise. +test_debug() { + test "$debug" = "" || eval "$1" +} + +# Public: Stop execution and start a shell. +# +# This is useful for debugging tests and only makes sense together with "-v". +# Be sure to remove all invocations of this command before submitting. +test_pause() { + if test "$verbose" = t; then + "$SHELL_PATH" <&6 >&3 2>&4 + else + error >&5 "test_pause requires --verbose" + fi +} + +# Public: Run test commands and expect them to succeed. +# +# When the test passed, an "ok" message is printed and the number of successful +# tests is incremented. When it failed, a "not ok" message is printed and the +# number of failed tests is incremented. +# +# With --immediate, exit test immediately upon the first failed test. +# +# Usually takes two arguments: +# $1 - Test description +# $2 - Commands to be executed. +# +# With three arguments, the first will be taken to be a prerequisite: +# $1 - Comma-separated list of test prerequisites. The test will be skipped if +# not all of the given prerequisites are set. To negate a prerequisite, +# put a "!" in front of it. +# $2 - Test description +# $3 - Commands to be executed. +# +# Examples +# +# test_expect_success \ +# 'git-write-tree should be able to write an empty tree.' \ +# 'tree=$(git-write-tree)' +# +# # Test depending on one prerequisite. +# test_expect_success TTY 'git --paginate rev-list uses a pager' \ +# ' ... ' +# +# # Multiple prerequisites are separated by a comma. +# test_expect_success PERL,PYTHON 'yo dawg' \ +# ' test $(perl -E 'print eval "1 +" . qx[python -c "print 2"]') == "4" ' +# +# Returns nothing. +test_expect_success() { + test "$#" = 3 && { test_prereq=$1; shift; } || test_prereq= + test "$#" = 2 || error "bug in the test script: not 2 or 3 parameters to test_expect_success" + export test_prereq + if ! test_skip_ "$@"; then + say >&3 "expecting success: $2" + if test_run_ "$2"; then + test_ok_ "$1" + else + test_failure_ "$@" + fi + fi + echo >&3 "" +} + +# Public: Run test commands and expect them to fail. Used to demonstrate a known +# breakage. +# +# This is NOT the opposite of test_expect_success, but rather used to mark a +# test that demonstrates a known breakage. +# +# When the test passed, an "ok" message is printed and the number of fixed tests +# is incremented. When it failed, a "not ok" message is printed and the number +# of tests still broken is incremented. +# +# Failures from these tests won't cause --immediate to stop. +# +# Usually takes two arguments: +# $1 - Test description +# $2 - Commands to be executed. +# +# With three arguments, the first will be taken to be a prerequisite: +# $1 - Comma-separated list of test prerequisites. The test will be skipped if +# not all of the given prerequisites are set. To negate a prerequisite, +# put a "!" in front of it. +# $2 - Test description +# $3 - Commands to be executed. +# +# Returns nothing. +test_expect_failure() { + test "$#" = 3 && { test_prereq=$1; shift; } || test_prereq= + test "$#" = 2 || error "bug in the test script: not 2 or 3 parameters to test_expect_failure" + export test_prereq + if ! test_skip_ "$@"; then + say >&3 "checking known breakage: $2" + if test_run_ "$2" expecting_failure; then + test_known_broken_ok_ "$1" + else + test_known_broken_failure_ "$1" + fi + fi + echo >&3 "" +} + +# Public: Run test commands and expect anything from them. Used when a +# test is not stable or not finished for some reason. +# +# When the test passed, an "ok" message is printed, but the number of +# fixed tests is not incremented. +# +# When it failed, a "not ok ... # TODO known breakage" message is +# printed, and the number of tests still broken is incremented. +# +# Failures from these tests won't cause --immediate to stop. +# +# Usually takes two arguments: +# $1 - Test description +# $2 - Commands to be executed. +# +# With three arguments, the first will be taken to be a prerequisite: +# $1 - Comma-separated list of test prerequisites. The test will be skipped if +# not all of the given prerequisites are set. To negate a prerequisite, +# put a "!" in front of it. +# $2 - Test description +# $3 - Commands to be executed. +# +# Returns nothing. +test_expect_unstable() { + test "$#" = 3 && { test_prereq=$1; shift; } || test_prereq= + test "$#" = 2 || error "bug in the test script: not 2 or 3 parameters to test_expect_unstable" + export test_prereq + if ! test_skip_ "$@"; then + say >&3 "checking unstable test: $2" + if test_run_ "$2" unstable; then + test_ok_ "$1" + else + test_known_broken_failure_ "$1" + fi + fi + echo >&3 "" +} + +# Public: Run command and ensure that it fails in a controlled way. +# +# Use it instead of "! ". For example, when dies due to a +# segfault, test_must_fail diagnoses it as an error, while "! " would +# mistakenly be treated as just another expected failure. +# +# This is one of the prefix functions to be used inside test_expect_success or +# test_expect_failure. +# +# $1.. - Command to be executed. +# +# Examples +# +# test_expect_success 'complain and die' ' +# do something && +# do something else && +# test_must_fail git checkout ../outerspace +# ' +# +# Returns 1 if the command succeeded (exit code 0). +# Returns 1 if the command died by signal (exit codes 130-192) +# Returns 1 if the command could not be found (exit code 127). +# Returns 0 otherwise. +test_must_fail() { + "$@" + exit_code=$? + if test $exit_code = 0; then + echo >&2 "test_must_fail: command succeeded: $*" + return 1 + elif test $exit_code -gt 129 -a $exit_code -le 192; then + echo >&2 "test_must_fail: died by signal: $*" + return 1 + elif test $exit_code = 127; then + echo >&2 "test_must_fail: command not found: $*" + return 1 + fi + return 0 +} + +# Public: Run command and ensure that it succeeds or fails in a controlled way. +# +# Similar to test_must_fail, but tolerates success too. Use it instead of +# " || :" to catch failures caused by a segfault, for instance. +# +# This is one of the prefix functions to be used inside test_expect_success or +# test_expect_failure. +# +# $1.. - Command to be executed. +# +# Examples +# +# test_expect_success 'some command works without configuration' ' +# test_might_fail git config --unset all.configuration && +# do something +# ' +# +# Returns 1 if the command died by signal (exit codes 130-192) +# Returns 1 if the command could not be found (exit code 127). +# Returns 0 otherwise. +test_might_fail() { + "$@" + exit_code=$? + if test $exit_code -gt 129 -a $exit_code -le 192; then + echo >&2 "test_might_fail: died by signal: $*" + return 1 + elif test $exit_code = 127; then + echo >&2 "test_might_fail: command not found: $*" + return 1 + fi + return 0 +} + +# Public: Run command and ensure it exits with a given exit code. +# +# This is one of the prefix functions to be used inside test_expect_success or +# test_expect_failure. +# +# $1 - Expected exit code. +# $2.. - Command to be executed. +# +# Examples +# +# test_expect_success 'Merge with d/f conflicts' ' +# test_expect_code 1 git merge "merge msg" B master +# ' +# +# Returns 0 if the expected exit code is returned or 1 otherwise. +test_expect_code() { + want_code=$1 + shift + "$@" + exit_code=$? + if test "$exit_code" = "$want_code"; then + return 0 + fi + + echo >&2 "test_expect_code: command exited with $exit_code, we wanted $want_code $*" + return 1 +} + +# Public: Compare two files to see if expected output matches actual output. +# +# The TEST_CMP variable defines the command used for the comparison; it +# defaults to "diff -u". Only when the test script was started with --verbose, +# will the command's output, the diff, be printed to the standard output. +# +# This is one of the prefix functions to be used inside test_expect_success or +# test_expect_failure. +# +# $1 - Path to file with expected output. +# $2 - Path to file with actual output. +# +# Examples +# +# test_expect_success 'foo works' ' +# echo expected >expected && +# foo >actual && +# test_cmp expected actual +# ' +# +# Returns the exit code of the command set by TEST_CMP. +test_cmp() { + ${TEST_CMP:-diff -u} "$@" +} + +# Public: portably print a sequence of numbers. +# +# seq is not in POSIX and GNU seq might not be available everywhere, +# so it is nice to have a seq implementation, even a very simple one. +# +# $1 - Starting number. +# $2 - Ending number. +# +# Examples +# +# test_expect_success 'foo works 10 times' ' +# for i in $(test_seq 1 10) +# do +# foo || return +# done +# ' +# +# Returns 0 if all the specified numbers can be displayed. +test_seq() { + i="$1" + j="$2" + while test "$i" -le "$j" + do + echo "$i" || return + i=$(("$i" + 1)) + done +} + +# Public: Check if the file expected to be empty is indeed empty, and barfs +# otherwise. +# +# $1 - File to check for emptiness. +# +# Returns 0 if file is empty, 1 otherwise. +test_must_be_empty() { + if test -s "$1" + then + echo "'$1' is not empty, it contains:" + cat "$1" + return 1 + fi +} + +# debugging-friendly alternatives to "test [-f|-d|-e]" +# The commands test the existence or non-existence of $1. $2 can be +# given to provide a more precise diagnosis. +test_path_is_file () { + if ! test -f "$1" + then + echo "File $1 doesn't exist. $2" + false + fi +} + +test_path_is_dir () { + if ! test -d "$1" + then + echo "Directory $1 doesn't exist. $2" + false + fi +} + +# Check if the directory exists and is empty as expected, barf otherwise. +test_dir_is_empty () { + test_path_is_dir "$1" && + if test -n "$(find "$1" -mindepth 1 -maxdepth 1)" + then + echo "Directory '$1' is not empty, it contains:" + ls -la "$1" + return 1 + fi +} + +# Public: Schedule cleanup commands to be run unconditionally at the end of a +# test. +# +# If some cleanup command fails, the test will not pass. With --immediate, no +# cleanup is done to help diagnose what went wrong. +# +# This is one of the prefix functions to be used inside test_expect_success or +# test_expect_failure. +# +# $1.. - Commands to prepend to the list of cleanup commands. +# +# Examples +# +# test_expect_success 'test core.capslock' ' +# git config core.capslock true && +# test_when_finished "git config --unset core.capslock" && +# do_something +# ' +# +# Returns the exit code of the last cleanup command executed. +test_when_finished() { + test_cleanup="{ $* + } && (exit \"\$eval_ret\"); eval_ret=\$?; $test_cleanup" +} + +# Public: Schedule cleanup commands to be run unconditionally when all tests +# have run. +# +# This can be used to clean up things like test databases. It is not needed to +# clean up temporary files, as test_done already does that. +# +# Examples: +# +# cleanup mysql -e "DROP DATABASE mytest" +# +# Returns the exit code of the last cleanup command executed. +final_cleanup= +cleanup() { + final_cleanup="{ $* + } && (exit \"\$eval_ret\"); eval_ret=\$?; $final_cleanup" +} + +# Public: Summarize test results and exit with an appropriate error code. +# +# Must be called at the end of each test script. +# +# Can also be used to stop tests early and skip all remaining tests. For this, +# set skip_all to a string explaining why the tests were skipped before calling +# test_done. +# +# Examples +# +# # Each test script must call test_done at the end. +# test_done +# +# # Skip all remaining tests if prerequisite is not set. +# if ! test_have_prereq PERL; then +# skip_all='skipping perl interface tests, perl not available' +# test_done +# fi +# +# Returns 0 if all tests passed or 1 if there was a failure. +test_done() { + EXIT_OK=t + + if test -z "$HARNESS_ACTIVE"; then + test_results_dir="$SHARNESS_TEST_OUTDIR/test-results" + mkdir -p "$test_results_dir" + test_results_path="$test_results_dir/$this_test.$$.counts" + + cat >>"$test_results_path" <<-EOF + total $SHARNESS_TEST_NB + success $test_success + fixed $test_fixed + broken $test_broken + failed $test_failure + + EOF + fi + + if test "$test_fixed" != 0; then + say_color error "# $test_fixed known breakage(s) vanished; please update test(s)" + fi + if test "$test_broken" != 0; then + say_color warn "# still have $test_broken known breakage(s)" + fi + if test "$test_broken" != 0 || test "$test_fixed" != 0; then + test_remaining=$((SHARNESS_TEST_NB - test_broken - test_fixed)) + msg="remaining $test_remaining test(s)" + else + test_remaining=$SHARNESS_TEST_NB + msg="$SHARNESS_TEST_NB test(s)" + fi + + case "$test_failure" in + 0) + # Maybe print SKIP message + check_skip_all_ + if test "$test_remaining" -gt 0; then + say_color pass "# passed all $msg" + fi + say "1..$SHARNESS_TEST_NB$skip_all" + + test_eval_ "$final_cleanup" + + remove_trash_ + + exit 0 ;; + + *) + say_color error "# failed $test_failure among $msg" + say "1..$SHARNESS_TEST_NB" + + exit 1 ;; + + esac +} diff --git a/test/sharness.sh b/test/sharness.sh index 1d57ce9..a3d8815 100644 --- a/test/sharness.sh +++ b/test/sharness.sh @@ -1,4 +1,4 @@ -#!/bin/sh +# Sharness test framework. # # Copyright (c) 2011-2012 Mathias Lafeldt # Copyright (c) 2005-2012 Git project @@ -18,14 +18,43 @@ # along with this program. If not, see http://www.gnu.org/licenses/ . # Public: Current version of Sharness. -SHARNESS_VERSION="1.0.0" +SHARNESS_VERSION="1.1.0" export SHARNESS_VERSION +: "${SHARNESS_TEST_EXTENSION:=t}" # Public: The file extension for tests. By default, it is set to "t". -: ${SHARNESS_TEST_EXTENSION:=t} export SHARNESS_TEST_EXTENSION -# Reset TERM to original terminal if found, otherwise save orignal TERM +if test -z "$SHARNESS_TEST_DIRECTORY" +then + SHARNESS_TEST_DIRECTORY=$(pwd) +else + # ensure that SHARNESS_TEST_DIRECTORY is an absolute path so that it + # is valid even if the current working directory is changed + SHARNESS_TEST_DIRECTORY=$(cd "$SHARNESS_TEST_DIRECTORY" && pwd) || exit 1 +fi +# Public: Root directory containing tests. Tests can override this variable, +# e.g. for testing Sharness itself. +export SHARNESS_TEST_DIRECTORY + +: "${SHARNESS_TEST_SRCDIR:=$(cd "$(dirname "$0")" && pwd)}" +# Public: Source directory of test code and sharness library. +# This directory may be different from the directory in which tests are +# being run. +export SHARNESS_TEST_SRCDIR + +if test -z "$SHARNESS_TEST_OUTDIR" +then + # Similarly, override this to store the test-results subdir + # elsewhere + SHARNESS_TEST_OUTDIR=$SHARNESS_TEST_DIRECTORY +fi + +# Public: Directory where the output of the tests should be stored (i.e. +# trash directories). +export SHARNESS_TEST_OUTDIR + +# Reset TERM to original terminal if found, otherwise save original TERM [ "x" = "x$SHARNESS_ORIG_TERM" ] && SHARNESS_ORIG_TERM="$TERM" || TERM="$SHARNESS_ORIG_TERM" @@ -33,23 +62,45 @@ export SHARNESS_TEST_EXTENSION export SHARNESS_ORIG_TERM # Export SHELL_PATH -: ${SHELL_PATH:=$SHELL} +: "${SHELL_PATH:=/bin/sh}" export SHELL_PATH +# if --tee was passed, write the output not only to the terminal, but +# additionally to the file test-results/$BASENAME.out, too. +case "$SHARNESS_TEST_TEE_STARTED, $* " in +done,*) + # do not redirect again + ;; +*' --tee '*|*' --verbose-log '*) + mkdir -p "$SHARNESS_TEST_OUTDIR/test-results" + BASE="$SHARNESS_TEST_OUTDIR/test-results/$(basename "$0" ".$SHARNESS_TEST_EXTENSION")" + + # Make this filename available to the sub-process in case it is using + # --verbose-log. + SHARNESS_TEST_TEE_OUTPUT_FILE="$BASE.out" + export SHARNESS_TEST_TEE_OUTPUT_FILE + + # Truncate before calling "tee -a" to get rid of the results + # from any previous runs. + : >"$SHARNESS_TEST_TEE_OUTPUT_FILE" + + (SHARNESS_TEST_TEE_STARTED="done" ${SHELL_PATH} "$0" "$@" 2>&1; + echo $? >"$BASE.exit") | tee -a "$SHARNESS_TEST_TEE_OUTPUT_FILE" + test "$(cat "$BASE.exit")" = 0 + exit + ;; +esac + # For repeatability, reset the environment to a known state. # TERM is sanitized below, after saving color control sequences. LANG=C LC_ALL=C -PAGER=cat +PAGER="cat" TZ=UTC EDITOR=: export LANG LC_ALL PAGER TZ EDITOR unset VISUAL CDPATH GREP_OPTIONS -# Line feed -LF=' -' - [ "x$TERM" != "xdumb" ] && ( [ -t 1 ] && tput bold >/dev/null 2>&1 && @@ -82,9 +133,17 @@ while test "$#" -ne 0; do chain_lint=; shift ;; --no-color) color=; shift ;; + --tee) + shift ;; # was handled already --root=*) root=$(expr "z$1" : 'z[^=]*=\(.*\)') shift ;; + -x) + trace=t + shift ;; + --verbose-log) + verbose_log=t + shift ;; *) echo "error: unknown test option '$1'" >&2; exit 1 ;; esac @@ -109,21 +168,62 @@ if test -n "$color"; then say_color_pass=$(tput setaf 2) # green say_color_info=$(tput setaf 6) # cyan say_color_reset=$(tput sgr0) - say_color_="" # no formatting for normal text + say_color_raw="" # no formatting for normal text say_color() { test -z "$1" && test -n "$quiet" && return - eval "say_color_color=\$say_color_$1" + case "$1" in + error) say_color_color=$say_color_error ;; + skip) say_color_color=$say_color_skip ;; + warn) say_color_color=$say_color_warn ;; + pass) say_color_color=$say_color_pass ;; + info) say_color_color=$say_color_info ;; + *) say_color_color=$say_color_raw ;; + esac shift - printf "%s\\n" "$say_color_color$*$say_color_reset" + printf '%s%s%s\n' "$say_color_color" "$*" "$say_color_reset" } else say_color() { test -z "$1" && test -n "$quiet" && return shift - printf "%s\n" "$*" + printf '%s\n' "$*" } fi +: "${test_untraceable:=}" +# Public: When set to a non-empty value, the current test will not be +# traced, unless it's run with a Bash version supporting +# BASH_XTRACEFD, i.e. v4.1 or later. +export test_untraceable + +if test -n "$trace" && test -n "$test_untraceable" +then + # '-x' tracing requested, but this test script can't be reliably + # traced, unless it is run with a Bash version supporting + # BASH_XTRACEFD (introduced in Bash v4.1). + # + # Perform this version check _after_ the test script was + # potentially re-executed with $TEST_SHELL_PATH for '--tee' or + # '--verbose-log', so the right shell is checked and the + # warning is issued only once. + if test -n "$BASH_VERSION" && eval ' + test ${BASH_VERSINFO[0]} -gt 4 || { + test ${BASH_VERSINFO[0]} -eq 4 && + test ${BASH_VERSINFO[1]} -ge 1 + } + ' + then + : Executed by a Bash version supporting BASH_XTRACEFD. Good. + else + echo >&2 "warning: ignoring -x; '$0' is untraceable without BASH_XTRACEFD" + trace= + fi +fi +if test -n "$trace" && test -z "$verbose_log" +then + verbose=t +fi + TERM=dumb export TERM @@ -137,7 +237,7 @@ say() { say_color info "$*" } -test -n "$test_description" || error "Test script did not set test_description." +test -n "${test_description:-}" || error "Test script did not set test_description." if test "$help" = "t"; then echo "$test_description" @@ -146,17 +246,32 @@ fi exec 5>&1 exec 6<&0 -if test "$verbose" = "t"; then +if test "$verbose_log" = "t" +then + exec 3>>"$SHARNESS_TEST_TEE_OUTPUT_FILE" 4>&3 +elif test "$verbose" = "t" +then exec 4>&2 3>&1 else exec 4>/dev/null 3>/dev/null fi -test_failure=0 -test_count=0 -test_fixed=0 -test_broken=0 -test_success=0 +# Send any "-x" output directly to stderr to avoid polluting tests +# which capture stderr. We can do this unconditionally since it +# has no effect if tracing isn't turned on. +# +# Note that this sets up the trace fd as soon as we assign the variable, so it +# must come after the creation of descriptor 4 above. Likewise, we must never +# unset this, as it has the side effect of closing descriptor 4, which we +# use to show verbose tests to the user. +# +# Note also that we don't need or want to export it. The tracing is local to +# this shell, and we would not want to influence any shells we exec. +BASH_XTRACEFD=4 + +# Public: The current test number, starting at 0. +SHARNESS_TEST_NB=0 +export SHARNESS_TEST_NB die() { code=$? @@ -171,159 +286,104 @@ die() { EXIT_OK= trap 'die' EXIT -# Public: Define that a test prerequisite is available. -# -# The prerequisite can later be checked explicitly using test_have_prereq or -# implicitly by specifying the prerequisite name in calls to test_expect_success -# or test_expect_failure. -# -# $1 - Name of prerequiste (a simple word, in all capital letters by convention) -# -# Examples -# -# # Set PYTHON prerequisite if interpreter is available. -# command -v python >/dev/null && test_set_prereq PYTHON -# -# # Set prerequisite depending on some variable. -# test -z "$NO_GETTEXT" && test_set_prereq GETTEXT -# -# Returns nothing. -test_set_prereq() { - satisfied_prereq="$satisfied_prereq$1 " -} -satisfied_prereq=" " - -# Public: Check if one or more test prerequisites are defined. -# -# The prerequisites must have previously been set with test_set_prereq. -# The most common use of this is to skip all the tests if some essential -# prerequisite is missing. -# -# $1 - Comma-separated list of test prerequisites. -# -# Examples -# -# # Skip all remaining tests if prerequisite is not set. -# if ! test_have_prereq PERL; then -# skip_all='skipping perl interface tests, perl not available' -# test_done -# fi -# -# Returns 0 if all prerequisites are defined or 1 otherwise. -test_have_prereq() { - # prerequisites can be concatenated with ',' - save_IFS=$IFS - IFS=, - set -- $* - IFS=$save_IFS - - total_prereq=0 - ok_prereq=0 - missing_prereq= - - for prerequisite; do - case "$prerequisite" in - !*) - negative_prereq=t - prerequisite=${prerequisite#!} - ;; - *) - negative_prereq= - esac - - total_prereq=$(($total_prereq + 1)) - case "$satisfied_prereq" in - *" $prerequisite "*) - satisfied_this_prereq=t - ;; - *) - satisfied_this_prereq= - esac +test_prereq= +missing_prereq= - case "$satisfied_this_prereq,$negative_prereq" in - t,|,t) - ok_prereq=$(($ok_prereq + 1)) - ;; - *) - # Keep a list of missing prerequisites; restore - # the negative marker if necessary. - prerequisite=${negative_prereq:+!}$prerequisite - if test -z "$missing_prereq"; then - missing_prereq=$prerequisite - else - missing_prereq="$prerequisite,$missing_prereq" - fi - esac - done +test_failure=0 +test_fixed=0 +test_broken=0 +test_success=0 - test $total_prereq = $ok_prereq -} +. "$SHARNESS_TEST_SRCDIR/lib-sharness/functions.sh" # You are not expected to call test_ok_ and test_failure_ directly, use # the text_expect_* functions instead. test_ok_() { - test_success=$(($test_success + 1)) - say_color "" "ok $test_count - $@" + test_success=$((test_success + 1)) + say_color "" "ok $SHARNESS_TEST_NB - $*" } test_failure_() { - test_failure=$(($test_failure + 1)) - say_color error "not ok $test_count - $1" + test_failure=$((test_failure + 1)) + say_color error "not ok $SHARNESS_TEST_NB - $1" shift echo "$@" | sed -e 's/^/# /' test "$immediate" = "" || { EXIT_OK=t; exit 1; } } test_known_broken_ok_() { - test_fixed=$(($test_fixed + 1)) - say_color error "ok $test_count - $@ # TODO known breakage vanished" + test_fixed=$((test_fixed + 1)) + say_color error "ok $SHARNESS_TEST_NB - $* # TODO known breakage vanished" } test_known_broken_failure_() { - test_broken=$(($test_broken + 1)) - say_color warn "not ok $test_count - $@ # TODO known breakage" + test_broken=$((test_broken + 1)) + say_color warn "not ok $SHARNESS_TEST_NB - $* # TODO known breakage" } -# Public: Execute commands in debug mode. -# -# Takes a single argument and evaluates it only when the test script is started -# with --debug. This is primarily meant for use during the development of test -# scripts. -# -# $1 - Commands to be executed. -# -# Examples -# -# test_debug "cat some_log_file" -# -# Returns the exit code of the last command executed in debug mode or 0 -# otherwise. -test_debug() { - test "$debug" = "" || eval "$1" +want_trace () { + test "$trace" = t && { + test "$verbose" = t || test "$verbose_log" = t + } } -# Public: Stop execution and start a shell. -# -# This is useful for debugging tests and only makes sense together with "-v". -# Be sure to remove all invocations of this command before submitting. -test_pause() { - if test "$verbose" = t; then - "$SHELL_PATH" <&6 >&3 2>&4 - else - error >&5 "test_pause requires --verbose" +# This is a separate function because some tests use +# "return" to end a test_expect_success block early +# (and we want to make sure we run any cleanup like +# "set +x"). +test_eval_inner_ () { + # Do not add anything extra (including LF) after '$*' + eval " + want_trace && set -x + $*" +} + +test_eval_x_ () { + # If "-x" tracing is in effect, then we want to avoid polluting stderr + # with non-test commands. But once in "set -x" mode, we cannot prevent + # the shell from printing the "set +x" to turn it off (nor the saving + # of $? before that). But we can make sure that the output goes to + # /dev/null. + # + # There are a few subtleties here: + # + # - we have to redirect descriptor 4 in addition to 2, to cover + # BASH_XTRACEFD + # + # - the actual eval has to come before the redirection block (since + # it needs to see descriptor 4 to set up its stderr) + # + # - likewise, any error message we print must be outside the block to + # access descriptor 4 + # + # - checking $? has to come immediately after the eval, but it must + # be _inside_ the block to avoid polluting the "set -x" output + # + + test_eval_inner_ "$@" &3 2>&4 + { + test_eval_ret_=$? + if want_trace + then + set +x + fi + } 2>/dev/null 4>&2 + + if test "$test_eval_ret_" != 0 && want_trace + then + say_color error >&4 "error: last command exited with \$?=$test_eval_ret_" fi + return $test_eval_ret_ } test_eval_() { - # This is a separate function because some tests use - # "return" to end a test_expect_success block early. case ",$test_prereq," in *,INTERACTIVE,*) eval "$*" ;; *) - eval &3 2>&4 "$*" + test_eval_x_ "$@" ;; esac } @@ -335,13 +395,22 @@ test_run_() { eval_ret=$? if test "$chain_lint" = "t"; then + # turn off tracing for this test-eval, as it simply creates + # confusing noise in the "-x" output + trace_tmp=$trace + trace= + # 117 is magic because it is unlikely to match the exit + # code of other programs test_eval_ "(exit 117) && $1" if test "$?" != 117; then error "bug in the test script: broken &&-chain: $1" fi + trace=$trace_tmp fi - if test -z "$immediate" || test $eval_ret = 0 || test -n "$expecting_failure"; then + if test -z "$immediate" || test $eval_ret = 0 || + test -n "$expecting_failure" && test "$test_cleanup" != ":" + then test_eval_ "$test_cleanup" fi if test "$verbose" = "t" && test -n "$HARNESS_ACTIVE"; then @@ -351,10 +420,10 @@ test_run_() { } test_skip_() { - test_count=$(($test_count + 1)) + SHARNESS_TEST_NB=$((SHARNESS_TEST_NB + 1)) to_skip= for skp in $SKIP_TESTS; do - case $this_test.$test_count in + case $this_test.$SHARNESS_TEST_NB in $skp) to_skip=t break @@ -370,8 +439,8 @@ test_skip_() { of_prereq=" of $test_prereq" fi - say_color skip >&3 "skipping test: $@" - say_color skip "ok $test_count # skip $1 (missing $missing_prereq${of_prereq})" + say_color skip >&3 "skipping test: $*" + say_color skip "ok $SHARNESS_TEST_NB # skip $1 (missing $missing_prereq${of_prereq})" : true ;; *) @@ -380,413 +449,30 @@ test_skip_() { esac } -# Public: Run test commands and expect them to succeed. -# -# When the test passed, an "ok" message is printed and the number of successful -# tests is incremented. When it failed, a "not ok" message is printed and the -# number of failed tests is incremented. -# -# With --immediate, exit test immediately upon the first failed test. -# -# Usually takes two arguments: -# $1 - Test description -# $2 - Commands to be executed. -# -# With three arguments, the first will be taken to be a prerequisite: -# $1 - Comma-separated list of test prerequisites. The test will be skipped if -# not all of the given prerequisites are set. To negate a prerequisite, -# put a "!" in front of it. -# $2 - Test description -# $3 - Commands to be executed. -# -# Examples -# -# test_expect_success \ -# 'git-write-tree should be able to write an empty tree.' \ -# 'tree=$(git-write-tree)' -# -# # Test depending on one prerequisite. -# test_expect_success TTY 'git --paginate rev-list uses a pager' \ -# ' ... ' -# -# # Multiple prerequisites are separated by a comma. -# test_expect_success PERL,PYTHON 'yo dawg' \ -# ' test $(perl -E 'print eval "1 +" . qx[python -c "print 2"]') == "4" ' -# -# Returns nothing. -test_expect_success() { - test "$#" = 3 && { test_prereq=$1; shift; } || test_prereq= - test "$#" = 2 || error "bug in the test script: not 2 or 3 parameters to test_expect_success" - export test_prereq - if ! test_skip_ "$@"; then - say >&3 "expecting success: $2" - if test_run_ "$2"; then - test_ok_ "$1" - else - test_failure_ "$@" - fi - fi - echo >&3 "" -} - -# Public: Run test commands and expect them to fail. Used to demonstrate a known -# breakage. -# -# This is NOT the opposite of test_expect_success, but rather used to mark a -# test that demonstrates a known breakage. -# -# When the test passed, an "ok" message is printed and the number of fixed tests -# is incremented. When it failed, a "not ok" message is printed and the number -# of tests still broken is incremented. -# -# Failures from these tests won't cause --immediate to stop. -# -# Usually takes two arguments: -# $1 - Test description -# $2 - Commands to be executed. -# -# With three arguments, the first will be taken to be a prerequisite: -# $1 - Comma-separated list of test prerequisites. The test will be skipped if -# not all of the given prerequisites are set. To negate a prerequisite, -# put a "!" in front of it. -# $2 - Test description -# $3 - Commands to be executed. -# -# Returns nothing. -test_expect_failure() { - test "$#" = 3 && { test_prereq=$1; shift; } || test_prereq= - test "$#" = 2 || error "bug in the test script: not 2 or 3 parameters to test_expect_failure" - export test_prereq - if ! test_skip_ "$@"; then - say >&3 "checking known breakage: $2" - if test_run_ "$2" expecting_failure; then - test_known_broken_ok_ "$1" - else - test_known_broken_failure_ "$1" - fi - fi - echo >&3 "" -} - -# Public: Run command and ensure that it fails in a controlled way. -# -# Use it instead of "! ". For example, when dies due to a -# segfault, test_must_fail diagnoses it as an error, while "! " would -# mistakenly be treated as just another expected failure. -# -# This is one of the prefix functions to be used inside test_expect_success or -# test_expect_failure. -# -# $1.. - Command to be executed. -# -# Examples -# -# test_expect_success 'complain and die' ' -# do something && -# do something else && -# test_must_fail git checkout ../outerspace -# ' -# -# Returns 1 if the command succeeded (exit code 0). -# Returns 1 if the command died by signal (exit codes 130-192) -# Returns 1 if the command could not be found (exit code 127). -# Returns 0 otherwise. -test_must_fail() { - "$@" - exit_code=$? - if test $exit_code = 0; then - echo >&2 "test_must_fail: command succeeded: $*" - return 1 - elif test $exit_code -gt 129 -a $exit_code -le 192; then - echo >&2 "test_must_fail: died by signal: $*" - return 1 - elif test $exit_code = 127; then - echo >&2 "test_must_fail: command not found: $*" - return 1 - fi - return 0 -} - -# Public: Run command and ensure that it succeeds or fails in a controlled way. -# -# Similar to test_must_fail, but tolerates success too. Use it instead of -# " || :" to catch failures caused by a segfault, for instance. -# -# This is one of the prefix functions to be used inside test_expect_success or -# test_expect_failure. -# -# $1.. - Command to be executed. -# -# Examples -# -# test_expect_success 'some command works without configuration' ' -# test_might_fail git config --unset all.configuration && -# do something -# ' -# -# Returns 1 if the command died by signal (exit codes 130-192) -# Returns 1 if the command could not be found (exit code 127). -# Returns 0 otherwise. -test_might_fail() { - "$@" - exit_code=$? - if test $exit_code -gt 129 -a $exit_code -le 192; then - echo >&2 "test_might_fail: died by signal: $*" - return 1 - elif test $exit_code = 127; then - echo >&2 "test_might_fail: command not found: $*" - return 1 - fi - return 0 -} - -# Public: Run command and ensure it exits with a given exit code. -# -# This is one of the prefix functions to be used inside test_expect_success or -# test_expect_failure. -# -# $1 - Expected exit code. -# $2.. - Command to be executed. -# -# Examples -# -# test_expect_success 'Merge with d/f conflicts' ' -# test_expect_code 1 git merge "merge msg" B master -# ' -# -# Returns 0 if the expected exit code is returned or 1 otherwise. -test_expect_code() { - want_code=$1 - shift - "$@" - exit_code=$? - if test $exit_code = $want_code; then - return 0 - fi - - echo >&2 "test_expect_code: command exited with $exit_code, we wanted $want_code $*" - return 1 -} - -# Public: Compare two files to see if expected output matches actual output. -# -# The TEST_CMP variable defines the command used for the comparision; it -# defaults to "diff -u". Only when the test script was started with --verbose, -# will the command's output, the diff, be printed to the standard output. -# -# This is one of the prefix functions to be used inside test_expect_success or -# test_expect_failure. -# -# $1 - Path to file with expected output. -# $2 - Path to file with actual output. -# -# Examples -# -# test_expect_success 'foo works' ' -# echo expected >expected && -# foo >actual && -# test_cmp expected actual -# ' -# -# Returns the exit code of the command set by TEST_CMP. -test_cmp() { - ${TEST_CMP:-diff -u} "$@" -} - -# Public: portably print a sequence of numbers. -# -# seq is not in POSIX and GNU seq might not be available everywhere, -# so it is nice to have a seq implementation, even a very simple one. -# -# $1 - Starting number. -# $2 - Ending number. -# -# Examples -# -# test_expect_success 'foo works 10 times' ' -# for i in $(test_seq 1 10) -# do -# foo || return -# done -# ' -# -# Returns 0 if all the specified numbers can be displayed. -test_seq() { - i="$1" - j="$2" - while test "$i" -le "$j" - do - echo "$i" || return - i=$(expr "$i" + 1) - done -} - -# Public: Check if the file expected to be empty is indeed empty, and barfs -# otherwise. -# -# $1 - File to check for emptyness. -# -# Returns 0 if file is empty, 1 otherwise. -test_must_be_empty() { - if test -s "$1" - then - echo "'$1' is not empty, it contains:" - cat "$1" - return 1 - fi -} - -# Public: Schedule cleanup commands to be run unconditionally at the end of a -# test. -# -# If some cleanup command fails, the test will not pass. With --immediate, no -# cleanup is done to help diagnose what went wrong. -# -# This is one of the prefix functions to be used inside test_expect_success or -# test_expect_failure. -# -# $1.. - Commands to prepend to the list of cleanup commands. -# -# Examples -# -# test_expect_success 'test core.capslock' ' -# git config core.capslock true && -# test_when_finished "git config --unset core.capslock" && -# do_something -# ' -# -# Returns the exit code of the last cleanup command executed. -test_when_finished() { - test_cleanup="{ $* - } && (exit \"\$eval_ret\"); eval_ret=\$?; $test_cleanup" -} - -# Public: Schedule cleanup commands to be run unconditionally when all tests -# have run. -# -# This can be used to clean up things like test databases. It is not needed to -# clean up temporary files, as test_done already does that. -# -# Examples: -# -# cleanup mysql -e "DROP DATABASE mytest" -# -# Returns the exit code of the last cleanup command executed. -final_cleanup= -cleanup() { - final_cleanup="{ $* - } && (exit \"\$eval_ret\"); eval_ret=\$?; $final_cleanup" -} - -# Public: Summarize test results and exit with an appropriate error code. -# -# Must be called at the end of each test script. -# -# Can also be used to stop tests early and skip all remaining tests. For this, -# set skip_all to a string explaining why the tests were skipped before calling -# test_done. -# -# Examples -# -# # Each test script must call test_done at the end. -# test_done -# -# # Skip all remaining tests if prerequisite is not set. -# if ! test_have_prereq PERL; then -# skip_all='skipping perl interface tests, perl not available' -# test_done -# fi -# -# Returns 0 if all tests passed or 1 if there was a failure. -test_done() { - EXIT_OK=t - - if test -z "$HARNESS_ACTIVE"; then - test_results_dir="$SHARNESS_TEST_DIRECTORY/test-results" - mkdir -p "$test_results_dir" - test_results_path="$test_results_dir/$this_test.$$.counts" - - cat >>"$test_results_path" <<-EOF - total $test_count - success $test_success - fixed $test_fixed - broken $test_broken - failed $test_failure - - EOF - fi - - if test "$test_fixed" != 0; then - say_color error "# $test_fixed known breakage(s) vanished; please update test(s)" - fi - if test "$test_broken" != 0; then - say_color warn "# still have $test_broken known breakage(s)" - fi - if test "$test_broken" != 0 || test "$test_fixed" != 0; then - test_remaining=$(( $test_count - $test_broken - $test_fixed )) - msg="remaining $test_remaining test(s)" - else - test_remaining=$test_count - msg="$test_count test(s)" - fi - - case "$test_failure" in - 0) - # Maybe print SKIP message - if test -n "$skip_all" && test $test_count -gt 0; then - error "Can't use skip_all after running some tests" - fi - [ -z "$skip_all" ] || skip_all=" # SKIP $skip_all" - - if test $test_remaining -gt 0; then - say_color pass "# passed all $msg" - fi - say "1..$test_count$skip_all" - - test_eval_ "$final_cleanup" - - test -d "$remove_trash" && - cd "$(dirname "$remove_trash")" && - rm -rf "$(basename "$remove_trash")" - - exit 0 ;; - - *) - say_color error "# failed $test_failure among $msg" - say "1..$test_count" - - exit 1 ;; - - esac -} - -# Public: Root directory containing tests. Tests can override this variable, -# e.g. for testing Sharness itself. -: ${SHARNESS_TEST_DIRECTORY:=$(pwd)} -export SHARNESS_TEST_DIRECTORY - -# Public: Source directory of test code and sharness library. -# This directory may be different from the directory in which tests are -# being run. -: ${SHARNESS_TEST_SRCDIR:=$(cd $(dirname $0) && pwd)} -export SHARNESS_TEST_SRCDIR - +: "${SHARNESS_BUILD_DIRECTORY:="$SHARNESS_TEST_DIRECTORY/.."}" # Public: Build directory that will be added to PATH. By default, it is set to # the parent directory of SHARNESS_TEST_DIRECTORY. -: ${SHARNESS_BUILD_DIRECTORY:="$SHARNESS_TEST_DIRECTORY/.."} +export SHARNESS_BUILD_DIRECTORY PATH="$SHARNESS_BUILD_DIRECTORY:$PATH" -export PATH SHARNESS_BUILD_DIRECTORY +export PATH # Public: Path to test script currently executed. SHARNESS_TEST_FILE="$0" export SHARNESS_TEST_FILE +remove_trash_() { + test -d "$remove_trash" && ( + cd "$(dirname "$remove_trash")" && + rm -rf "$(basename "$remove_trash")" + ) +} + # Prepare test area. SHARNESS_TRASH_DIRECTORY="trash directory.$(basename "$SHARNESS_TEST_FILE" ".$SHARNESS_TEST_EXTENSION")" test -n "$root" && SHARNESS_TRASH_DIRECTORY="$root/$SHARNESS_TRASH_DIRECTORY" case "$SHARNESS_TRASH_DIRECTORY" in /*) ;; # absolute path is good - *) SHARNESS_TRASH_DIRECTORY="$SHARNESS_TEST_DIRECTORY/$SHARNESS_TRASH_DIRECTORY" ;; + *) SHARNESS_TRASH_DIRECTORY="$SHARNESS_TEST_OUTDIR/$SHARNESS_TRASH_DIRECTORY" ;; esac test "$debug" = "t" || remove_trash="$SHARNESS_TRASH_DIRECTORY" rm -rf "$SHARNESS_TRASH_DIRECTORY" || { @@ -831,6 +517,13 @@ mkdir -p "$SHARNESS_TRASH_DIRECTORY" || exit 1 # in subprocesses like git equals our $PWD (for pathname comparisons). cd -P "$SHARNESS_TRASH_DIRECTORY" || exit 1 +check_skip_all_() { + if test -n "$skip_all" && test $SHARNESS_TEST_NB -gt 0; then + error "Can't use skip_all after running some tests" + fi + [ -z "$skip_all" ] || skip_all=" # SKIP $skip_all" +} + this_test=${SHARNESS_TEST_FILE##*/} this_test=${this_test%.$SHARNESS_TEST_EXTENSION} for skp in $SKIP_TESTS; do -- cgit v1.2.3