diff options
| author | Garrett D'Amore <garrett@damore.org> | 2020-11-21 22:11:21 -0800 |
|---|---|---|
| committer | Garrett D'Amore <garrett@damore.org> | 2020-11-23 22:20:12 -0800 |
| commit | d1218d7309475193b53911667911c4f59a1a7752 (patch) | |
| tree | 6ea796998fb60d2cb8afa704faa77fe7fddd644c /tests/acutest.h | |
| parent | b826bfc171d90f8bde7bd672c0ac14201b8b2742 (diff) | |
| download | nng-d1218d7309475193b53911667911c4f59a1a7752.tar.gz nng-d1218d7309475193b53911667911c4f59a1a7752.tar.bz2 nng-d1218d7309475193b53911667911c4f59a1a7752.zip | |
New NUTS test framework (NNG Unit Test Support).
This is based on testutil/acutest, but is cleaner and fixes some
short-comings. We will be adding more support for additional
common paradigms to better facilitate transport tests.
While here we added some more test cases, and fixed a possible
symbol collision in the the stats framework (due to Linux use
of a macro definition of "si_value" in a standard OS header).
Test coverage may regress slightly as we are no longer using
some of the legacy APIs.
Diffstat (limited to 'tests/acutest.h')
| -rw-r--r-- | tests/acutest.h | 1645 |
1 files changed, 0 insertions, 1645 deletions
diff --git a/tests/acutest.h b/tests/acutest.h deleted file mode 100644 index 7f0fc059..00000000 --- a/tests/acutest.h +++ /dev/null @@ -1,1645 +0,0 @@ -/* - * Acutest -- Another C/C++ Unit Test facility - * <http://github.com/mity/acutest> - * - * Copyright (c) 2013-2019 Martin Mitas - * - * Permission is hereby granted, free of charge, to any person obtaining a - * copy of this software and associated documentation files (the "Software"), - * to deal in the Software without restriction, including without limitation - * the rights to use, copy, modify, merge, publish, distribute, sublicense, - * and/or sell copies of the Software, and to permit persons to whom the - * Software is furnished to do so, subject to the following conditions: - * - * The above copyright notice and this permission notice shall be included in - * all copies or substantial portions of the Software. - * - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS - * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE - * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER - * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING - * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS - * IN THE SOFTWARE. - */ - -#ifndef ACUTEST_H__ -#define ACUTEST_H__ - - -/************************ - *** Public interface *** - ************************/ - -/* By default, "acutest.h" provides the main program entry point (function - * main()). However, if the test suite is composed of multiple source files - * which include "acutest.h", then this causes a problem of multiple main() - * definitions. To avoid this problem, #define macro TEST_NO_MAIN in all - * compilation units but one. - */ - -/* Macro to specify list of unit tests in the suite. - * The unit test implementation MUST provide list of unit tests it implements - * with this macro: - * - * TEST_LIST = { - * { "test1_name", test1_func_ptr }, - * { "test2_name", test2_func_ptr }, - * ... - * { 0 } - * }; - * - * The list specifies names of each test (must be unique) and pointer to - * a function implementing it. The function does not take any arguments - * and has no return values, i.e. every test function has to be compatible - * with this prototype: - * - * void test_func(void); - */ -#define TEST_LIST const struct test__ test_list__[] - - -/* Macros for testing whether an unit test succeeds or fails. These macros - * can be used arbitrarily in functions implementing the unit tests. - * - * If any condition fails throughout execution of a test, the test fails. - * - * TEST_CHECK takes only one argument (the condition), TEST_CHECK_ allows - * also to specify an error message to print out if the condition fails. - * (It expects printf-like format string and its parameters). The macros - * return non-zero (condition passes) or 0 (condition fails). - * - * That can be useful when more conditions should be checked only if some - * preceding condition passes, as illustrated in this code snippet: - * - * SomeStruct* ptr = allocate_some_struct(); - * if(TEST_CHECK(ptr != NULL)) { - * TEST_CHECK(ptr->member1 < 100); - * TEST_CHECK(ptr->member2 > 200); - * } - */ -#define TEST_CHECK_(cond,...) test_check__((cond), __FILE__, __LINE__, __VA_ARGS__) -#define TEST_CHECK(cond) test_check__((cond), __FILE__, __LINE__, "%s", #cond) - - -/* These macros are the same as TEST_CHECK_ and TEST_CHECK except that if the - * condition fails, the currently executed unit test is immediately aborted. - * - * That is done either by calling abort() if the unit test is executed as a - * child process; or via longjmp() if the unit test is executed within the - * main Acutest process. - * - * As a side effect of such abortion, your unit tests may cause memory leaks, - * unflushed file descriptors, and other fenomena caused by the abortion. - * - * Therefore you should not use these as a general replacement for TEST_CHECK. - * Use it with some caution, especially if your test causes some other side - * effects to the outside world (e.g. communicating with some server, inserting - * into a database etc.). - */ -#define TEST_ASSERT_(cond,...) \ - do { \ - if (!test_check__((cond), __FILE__, __LINE__, __VA_ARGS__)) \ - test_abort__(); \ - } while(0) -#define TEST_ASSERT(cond) \ - do { \ - if (!test_check__((cond), __FILE__, __LINE__, "%s", #cond)) \ - test_abort__(); \ - } while(0) - - -#ifdef __cplusplus -/* Macros to verify that the code (the 1st argument) throws exception of given - * type (the 2nd argument). (Note these macros are only available in C++.) - * - * TEST_EXCEPTION_ is like TEST_EXCEPTION but accepts custom printf-like - * message. - * - * For example: - * - * TEST_EXCEPTION(function_that_throw(), ExpectedExceptionType); - * - * If the function_that_throw() throws ExpectedExceptionType, the check passes. - * If the function throws anything incompatible with ExpectedExceptionType - * (or if it does not thrown an exception at all), the check fails. - */ -#define TEST_EXCEPTION(code, exctype) \ - do { \ - bool exc_ok__ = false; \ - const char *msg__ = NULL; \ - try { \ - code; \ - msg__ = "No exception thrown."; \ - } catch(exctype const&) { \ - exc_ok__= true; \ - } catch(...) { \ - msg__ = "Unexpected exception thrown."; \ - } \ - test_check__(exc_ok__, __FILE__, __LINE__, #code " throws " #exctype); \ - if(msg__ != NULL) \ - test_message__("%s", msg__); \ - } while(0) -#define TEST_EXCEPTION_(code, exctype, ...) \ - do { \ - bool exc_ok__ = false; \ - const char *msg__ = NULL; \ - try { \ - code; \ - msg__ = "No exception thrown."; \ - } catch(exctype const&) { \ - exc_ok__= true; \ - } catch(...) { \ - msg__ = "Unexpected exception thrown."; \ - } \ - test_check__(exc_ok__, __FILE__, __LINE__, __VA_ARGS__); \ - if(msg__ != NULL) \ - test_message__("%s", msg__); \ - } while(0) -#endif /* #ifdef __cplusplus */ - - -/* Sometimes it is useful to split execution of more complex unit tests to some - * smaller parts and associate those parts with some names. - * - * This is especially handy if the given unit test is implemented as a loop - * over some vector of multiple testing inputs. Using these macros allow to use - * sort of subtitle for each iteration of the loop (e.g. outputting the input - * itself or a name associated to it), so that if any TEST_CHECK condition - * fails in the loop, it can be easily seen which iteration triggers the - * failure, without the need to manually output the iteration-specific data in - * every single TEST_CHECK inside the loop body. - * - * TEST_CASE allows to specify only single string as the name of the case, - * TEST_CASE_ provides all the power of printf-like string formatting. - * - * Note that the test cases cannot be nested. Starting a new test case ends - * implicitly the previous one. To end the test case explicitly (e.g. to end - * the last test case after exiting the loop), you may use TEST_CASE(NULL). - */ -#define TEST_CASE_(...) test_case__(__VA_ARGS__) -#define TEST_CASE(name) test_case__("%s", name) - - -/* printf-like macro for outputting an extra information about a failure. - * - * Intended use is to output some computed output versus the expected value, - * e.g. like this: - * - * if(!TEST_CHECK(produced == expected)) { - * TEST_MSG("Expected: %d", expected); - * TEST_MSG("Produced: %d", produced); - * } - * - * Note the message is only written down if the most recent use of any checking - * macro (like e.g. TEST_CHECK or TEST_EXCEPTION) in the current test failed. - * This means the above is equivalent to just this: - * - * TEST_CHECK(produced == expected); - * TEST_MSG("Expected: %d", expected); - * TEST_MSG("Produced: %d", produced); - * - * The macro can deal with multi-line output fairly well. It also automatically - * adds a final new-line if there is none present. - */ -#define TEST_MSG(...) test_message__(__VA_ARGS__) - - -/* Maximal output per TEST_MSG call. Longer messages are cut. - * You may define another limit prior including "acutest.h" - */ -#ifndef TEST_MSG_MAXSIZE - #define TEST_MSG_MAXSIZE 1024 -#endif - - -/* Macro for dumping a block of memory. - * - * Its intended use is very similar to what TEST_MSG is for, but instead of - * generating any printf-like message, this is for dumping raw block of a - * memory in a hexadecimal form: - * - * TEST_CHECK(size_produced == size_expected && memcmp(addr_produced, addr_expected, size_produced) == 0); - * TEST_DUMP("Expected:", addr_expected, size_expected); - * TEST_DUMP("Produced:", addr_produced, size_produced); - */ -#define TEST_DUMP(title, addr, size) test_dump__(title, addr, size) - -/* Maximal output per TEST_DUMP call (in bytes to dump). Longer blocks are cut. - * You may define another limit prior including "acutest.h" - */ -#ifndef TEST_DUMP_MAXSIZE - #define TEST_DUMP_MAXSIZE 1024 -#endif - - -/********************** - *** Implementation *** - **********************/ - -/* The unit test files should not rely on anything below. */ - -#include <ctype.h> -#include <stdarg.h> -#include <stdio.h> -#include <stdlib.h> -#include <string.h> -#include <setjmp.h> - -#if defined(unix) || defined(__unix__) || defined(__unix) || defined(__APPLE__) - #define ACUTEST_UNIX__ 1 - #include <errno.h> - #include <libgen.h> - #include <unistd.h> - #include <sys/types.h> - #include <sys/wait.h> - #include <signal.h> - #include <time.h> - - #if defined CLOCK_PROCESS_CPUTIME_ID && defined CLOCK_MONOTONIC - #define ACUTEST_HAS_POSIX_TIMER__ 1 - #endif -#endif - -#if defined(__gnu_linux__) - #define ACUTEST_LINUX__ 1 - #include <fcntl.h> - #include <sys/stat.h> -#endif - -#if defined(_WIN32) || defined(__WIN32__) || defined(__WINDOWS__) - #define ACUTEST_WIN__ 1 - #include <windows.h> - #include <io.h> -#endif - -#ifdef __cplusplus - #include <exception> -#endif - - -/* Note our global private identifiers end with '__' to mitigate risk of clash - * with the unit tests implementation. */ - - -#ifdef __cplusplus - extern "C" { -#endif - - -struct test__ { - const char* name; - void (*func)(void); -}; - -struct test_detail__ { - unsigned char flags; - double duration; -}; - -enum { - TEST_FLAG_RUN__ = 1 << 0, - TEST_FLAG_SUCCESS__ = 1 << 1, - TEST_FLAG_FAILURE__ = 1 << 2, -}; - -extern const struct test__ test_list__[]; - -int test_check__(int cond, const char* file, int line, const char* fmt, ...); -void test_case__(const char* fmt, ...); -void test_message__(const char* fmt, ...); -void test_dump__(const char* title, const void* addr, size_t size); -void test_abort__(void); - - -#ifndef TEST_NO_MAIN - -static char* test_argv0__ = NULL; -static size_t test_list_size__ = 0; -static struct test_detail__ *test_details__ = NULL; -static size_t test_count__ = 0; -static int test_no_exec__ = -1; -static int test_no_summary__ = 0; -static int test_tap__ = 0; -static int test_skip_mode__ = 0; -static int test_worker__ = 0; -static int test_worker_index__ = 0; -static int test_cond_failed__ = 0; -static FILE *test_xml_output__ = NULL; - -static int test_stat_failed_units__ = 0; -static int test_stat_run_units__ = 0; - -static const struct test__* test_current_unit__ = NULL; -static int test_current_index__ = 0; -static char test_case_name__[64] = ""; -static int test_current_already_logged__ = 0; -static int test_case_current_already_logged__ = 0; -static int test_verbose_level__ = 2; -static int test_current_failures__ = 0; -static int test_colorize__ = 0; -static int test_timer__ = 0; - -static int test_abort_has_jmp_buf__ = 0; -static jmp_buf test_abort_jmp_buf__; - -#if defined ACUTEST_WIN__ - typedef LARGE_INTEGER test_timer_type__; - static LARGE_INTEGER test_timer_freq__; - static test_timer_type__ test_timer_start__; - static test_timer_type__ test_timer_end__; - - static void - test_timer_init__(void) - { - QueryPerformanceFrequency(&test_timer_freq__); - } - - static void - test_timer_get_time__(LARGE_INTEGER* ts) - { - QueryPerformanceCounter(ts); - } - - static double - test_timer_diff__(LARGE_INTEGER start, LARGE_INTEGER end) - { - double duration = (double)(end.QuadPart - start.QuadPart); - duration /= (double)test_timer_freq__.QuadPart; - return duration; - } - - static void - test_timer_print_diff__(void) - { - printf("%.6lf secs", test_timer_diff__(test_timer_start__, test_timer_end__)); - } -#elif defined ACUTEST_HAS_POSIX_TIMER__ - static clockid_t test_timer_id__; - typedef struct timespec test_timer_type__; - static test_timer_type__ test_timer_start__; - static test_timer_type__ test_timer_end__; - - static void - test_timer_init__(void) - { - if(test_timer__ == 1) - test_timer_id__ = CLOCK_MONOTONIC; - else if(test_timer__ == 2) - test_timer_id__ = CLOCK_PROCESS_CPUTIME_ID; - } - - static void - test_timer_get_time__(struct timespec* ts) - { - clock_gettime(test_timer_id__, ts); - } - - static double - test_timer_diff__(struct timespec start, struct timespec end) - { - double endns; - double startns; - - endns = end.tv_sec; - endns *= 1e9; - endns += end.tv_nsec; - - startns = start.tv_sec; - startns *= 1e9; - startns += start.tv_nsec; - - return ((endns - startns)/ 1e9); - } - - static void - test_timer_print_diff__(void) - { - printf("%.6lf secs", - test_timer_diff__(test_timer_start__, test_timer_end__)); - } -#else - typedef int test_timer_type__; - static test_timer_type__ test_timer_start__; - static test_timer_type__ test_timer_end__; - - void - test_timer_init__(void) - {} - - static void - test_timer_get_time__(int* ts) - { - (void) ts; - } - - static double - test_timer_diff__(int start, int end) - { - (void) start; - (void) end; - return 0.0; - } - - static void - test_timer_print_diff__(void) - {} -#endif - -#define TEST_COLOR_DEFAULT__ 0 -#define TEST_COLOR_GREEN__ 1 -#define TEST_COLOR_RED__ 2 -#define TEST_COLOR_DEFAULT_INTENSIVE__ 3 -#define TEST_COLOR_GREEN_INTENSIVE__ 4 -#define TEST_COLOR_RED_INTENSIVE__ 5 - -static int -test_print_in_color__(int color, const char* fmt, ...) -{ - va_list args; - char buffer[256]; - int n; - - va_start(args, fmt); - vsnprintf(buffer, sizeof(buffer), fmt, args); - va_end(args); - buffer[sizeof(buffer)-1] = '\0'; - - if(!test_colorize__) { - return printf("%s", buffer); - } - -#if defined ACUTEST_UNIX__ - { - const char* col_str; - switch(color) { - case TEST_COLOR_GREEN__: col_str = "\033[0;32m"; break; - case TEST_COLOR_RED__: col_str = "\033[0;31m"; break; - case TEST_COLOR_GREEN_INTENSIVE__: col_str = "\033[1;32m"; break; - case TEST_COLOR_RED_INTENSIVE__: col_str = "\033[1;31m"; break; - case TEST_COLOR_DEFAULT_INTENSIVE__: col_str = "\033[1m"; break; - default: col_str = "\033[0m"; break; - } - printf("%s", col_str); - n = printf("%s", buffer); - printf("\033[0m"); - return n; - } -#elif defined ACUTEST_WIN__ - { - HANDLE h; - CONSOLE_SCREEN_BUFFER_INFO info; - WORD attr; - - h = GetStdHandle(STD_OUTPUT_HANDLE); - GetConsoleScreenBufferInfo(h, &info); - - switch(color) { - case TEST_COLOR_GREEN__: attr = FOREGROUND_GREEN; break; - case TEST_COLOR_RED__: attr = FOREGROUND_RED; break; - case TEST_COLOR_GREEN_INTENSIVE__: attr = FOREGROUND_GREEN | FOREGROUND_INTENSITY; break; - case TEST_COLOR_RED_INTENSIVE__: attr = FOREGROUND_RED | FOREGROUND_INTENSITY; break; - case TEST_COLOR_DEFAULT_INTENSIVE__: attr = FOREGROUND_BLUE | FOREGROUND_GREEN | FOREGROUND_RED | FOREGROUND_INTENSITY; break; - default: attr = 0; break; - } - if(attr != 0) - SetConsoleTextAttribute(h, attr); - n = printf("%s", buffer); - SetConsoleTextAttribute(h, info.wAttributes); - return n; - } -#else - n = printf("%s", buffer); - return n; -#endif -} - -static void -test_begin_test_line__(const struct test__* test) -{ - if(!test_tap__) { - if(test_verbose_level__ >= 3) { - test_print_in_color__(TEST_COLOR_DEFAULT_INTENSIVE__, "Test %s:\n", test->name); - test_current_already_logged__++; - } else if(test_verbose_level__ >= 1) { - int n; - char spaces[48]; - - n = test_print_in_color__(TEST_COLOR_DEFAULT_INTENSIVE__, "Test %s... ", test->name); - memset(spaces, ' ', sizeof(spaces)); - if(n < (int) sizeof(spaces)) - printf("%.*s", (int) sizeof(spaces) - n, spaces); - } else { - test_current_already_logged__ = 1; - } - } -} - -static void -test_finish_test_line__(int result) -{ - if(test_tap__) { - const char* str = (result == 0) ? "ok" : "not ok"; - - printf("%s %u - %s\n", str, test_current_index__ + 1, test_current_unit__->name); - - if(result == 0 && test_timer__) { - printf("# Duration: "); - test_timer_print_diff__(); - printf("\n"); - } - } else { - int color = (result == 0) ? TEST_COLOR_GREEN_INTENSIVE__ : TEST_COLOR_RED_INTENSIVE__; - const char* str = (result == 0) ? "OK" : "FAILED"; - printf("[ "); - test_print_in_color__(color, str); - printf(" ]"); - - if(result == 0 && test_timer__) { - printf(" "); - test_timer_print_diff__(); - } - - printf("\n"); - } -} - -static void -test_line_indent__(int level) -{ - static const char spaces[] = " "; - int n = level * 2; - - if(test_tap__ && n > 0) { - n--; - printf("#"); - } - - while(n > 16) { - printf("%s", spaces); - n -= 16; - } - printf("%.*s", n, spaces); -} - -int -test_check__(int cond, const char* file, int line, const char* fmt, ...) -{ - const char *result_str; - int result_color; - int verbose_level; - - if(cond) { - result_str = "ok"; - result_color = TEST_COLOR_GREEN__; - verbose_level = 3; - } else { - if(!test_current_already_logged__ && test_current_unit__ != NULL) - test_finish_test_line__(-1); - - result_str = "failed"; - result_color = TEST_COLOR_RED__; - verbose_level = 2; - test_current_failures__++; - test_current_already_logged__++; - } - - if(test_verbose_level__ >= verbose_level) { - va_list args; - - if(!test_case_current_already_logged__ && test_case_name__[0]) { - test_line_indent__(1); - test_print_in_color__(TEST_COLOR_DEFAULT_INTENSIVE__, "Case %s:\n", test_case_name__); - test_current_already_logged__++; - test_case_current_already_logged__++; - } - - test_line_indent__(test_case_name__[0] ? 2 : 1); - if(file != NULL) { - if(test_verbose_level__ < 3) { -#ifdef ACUTEST_WIN__ - const char* lastsep1 = strrchr(file, '\\'); - const char* lastsep2 = strrchr(file, '/'); - if(lastsep1 == NULL) - lastsep1 = file-1; - if(lastsep2 == NULL) - lastsep2 = file-1; - file = (lastsep1 > lastsep2 ? lastsep1 : lastsep2) + 1; -#else - const char* lastsep = strrchr(file, '/'); - if(lastsep != NULL) - file = lastsep+1; -#endif - } - printf("%s:%d: Check ", file, line); - } - - va_start(args, fmt); - vprintf(fmt, args); - va_end(args); - - printf("... "); - test_print_in_color__(result_color, result_str); - printf("\n"); - test_current_already_logged__++; - } - - test_cond_failed__ = (cond == 0); - return !test_cond_failed__; -} - -void -test_case__(const char* fmt, ...) -{ - va_list args; - - if(test_verbose_level__ < 2) - return; - - if(test_case_name__[0]) { - test_case_current_already_logged__ = 0; - test_case_name__[0] = '\0'; - } - - if(fmt == NULL) - return; - - va_start(args, fmt); - vsnprintf(test_case_name__, sizeof(test_case_name__) - 1, fmt, args); - va_end(args); - test_case_name__[sizeof(test_case_name__) - 1] = '\0'; - - if(test_verbose_level__ >= 3) { - test_line_indent__(1); - test_print_in_color__(TEST_COLOR_DEFAULT_INTENSIVE__, "Case %s:\n", test_case_name__); - test_current_already_logged__++; - test_case_current_already_logged__++; - } -} - -void -test_message__(const char* fmt, ...) -{ - char buffer[TEST_MSG_MAXSIZE]; - char* line_beg; - char* line_end; - va_list args; - - if(test_verbose_level__ < 2) - return; - - /* We allow extra message only when something is already wrong in the - * current test. */ - if(test_current_unit__ == NULL || !test_cond_failed__) - return; - - va_start(args, fmt); - vsnprintf(buffer, TEST_MSG_MAXSIZE, fmt, args); - va_end(args); - buffer[TEST_MSG_MAXSIZE-1] = '\0'; - - line_beg = buffer; - while(1) { - line_end = strchr(line_beg, '\n'); - if(line_end == NULL) - break; - test_line_indent__(test_case_name__[0] ? 3 : 2); - printf("%.*s\n", (int)(line_end - line_beg), line_beg); - line_beg = line_end + 1; - } - if(line_beg[0] != '\0') { - test_line_indent__(test_case_name__[0] ? 3 : 2); - printf("%s\n", line_beg); - } -} - -void -test_dump__(const char* title, const void* addr, size_t size) -{ - static const size_t BYTES_PER_LINE = 16; - size_t line_beg; - size_t truncate = 0; - - if(test_verbose_level__ < 2) - return; - - /* We allow extra message only when something is already wrong in the - * current test. */ - if(test_current_unit__ == NULL || !test_cond_failed__) - return; - - if(size > TEST_DUMP_MAXSIZE) { - truncate = size - TEST_DUMP_MAXSIZE; - size = TEST_DUMP_MAXSIZE; - } - - test_line_indent__(test_case_name__[0] ? 3 : 2); - printf((title[strlen(title)-1] == ':') ? "%s\n" : "%s:\n", title); - - for(line_beg = 0; line_beg < size; line_beg += BYTES_PER_LINE) { - size_t line_end = line_beg + BYTES_PER_LINE; - size_t off; - - test_line_indent__(test_case_name__[0] ? 4 : 3); - printf("%08lx: ", (unsigned long)line_beg); - for(off = line_beg; off < line_end; off++) { - if(off < size) - printf(" %02x", ((unsigned char*)addr)[off]); - else - printf(" "); - } - - printf(" "); - for(off = line_beg; off < line_end; off++) { - unsigned char byte = ((unsigned char*)addr)[off]; - if(off < size) - printf("%c", (iscntrl(byte) ? '.' : byte)); - else - break; - } - - printf("\n"); - } - - if(truncate > 0) { - test_line_indent__(test_case_name__[0] ? 4 : 3); - printf(" ... (and more %u bytes)\n", (unsigned) truncate); - } -} - -void -test_abort__(void) -{ - if(test_abort_has_jmp_buf__) - longjmp(test_abort_jmp_buf__, 1); - else - abort(); -} - -static void -test_list_names__(void) -{ - const struct test__* test; - - printf("Unit tests:\n"); - for(test = &test_list__[0]; test->func != NULL; test++) - printf(" %s\n", test->name); -} - -static void -test_remember__(int i) -{ - if(test_details__[i].flags & TEST_FLAG_RUN__) - return; - - test_details__[i].flags |= TEST_FLAG_RUN__; - test_count__++; -} - -static void -test_set_success__(int i, int success) -{ - test_details__[i].flags |= success ? TEST_FLAG_SUCCESS__ : TEST_FLAG_FAILURE__; -} - -static void -test_set_duration__(int i, double duration) -{ - test_details__[i].duration = duration; -} - -static int -test_name_contains_word__(const char* name, const char* pattern) -{ - static const char word_delim[] = " \t-_."; - const char* substr; - size_t pattern_len; - int starts_on_word_boundary; - int ends_on_word_boundary; - - pattern_len = strlen(pattern); - - substr = strstr(name, pattern); - while(substr != NULL) { - starts_on_word_boundary = (substr == name || strchr(word_delim, substr[-1]) != NULL); - ends_on_word_boundary = (substr[pattern_len] == '\0' || strchr(word_delim, substr[pattern_len]) != NULL); - - if(starts_on_word_boundary && ends_on_word_boundary) - return 1; - - substr = strstr(substr+1, pattern); - } - - return 0; -} - -static int -test_lookup__(const char* pattern) -{ - int i; - int n = 0; - - /* Try exact match. */ - for(i = 0; i < (int) test_list_size__; i++) { - if(strcmp(test_list__[i].name, pattern) == 0) { - test_remember__(i); - n++; - break; - } - } - if(n > 0) - return n; - - /* Try word match. */ - for(i = 0; i < (int) test_list_size__; i++) { - if(test_name_contains_word__(test_list__[i].name, pattern)) { - test_remember__(i); - n++; - } - } - if(n > 0) - return n; - - /* Try relaxed match. */ - for(i = 0; i < (int) test_list_size__; i++) { - if(strstr(test_list__[i].name, pattern) != NULL) { - test_remember__(i); - n++; - } - } - - return n; -} - - -/* Called if anything goes bad in Acutest, or if the unit test ends in other - * way then by normal returning from its function (e.g. exception or some - * abnormal child process termination). */ -static void -test_error__(const char* fmt, ...) -{ - va_list args; - - if(test_verbose_level__ == 0) - return; - - if(test_verbose_level__ <= 2 && !test_current_already_logged__ && test_current_unit__ != NULL) { - if(test_tap__) { - test_finish_test_line__(-1); - } else { - printf("[ "); - test_print_in_color__(TEST_COLOR_RED_INTENSIVE__, "FAILED"); - printf(" ]\n"); - } - } - - if(test_verbose_level__ >= 2) { - test_line_indent__(1); - if(test_verbose_level__ >= 3) - test_print_in_color__(TEST_COLOR_RED_INTENSIVE__, "ERROR: "); - va_start(args, fmt); - vprintf(fmt, args); - va_end(args); - printf("\n"); - } - - if(test_verbose_level__ >= 3) { - printf("\n"); - } -} - -/* Call directly the given test unit function. */ -static int -test_do_run__(const struct test__* test, int index) -{ - test_current_unit__ = test; - test_current_index__ = index; - test_current_failures__ = 0; - test_current_already_logged__ = 0; - test_cond_failed__ = 0; - - test_begin_test_line__(test); - -#ifdef __cplusplus - try { -#endif - - /* This is good to do for case the test unit e.g. crashes. */ - fflush(stdout); - fflush(stderr); - - if(!test_worker__) { - test_abort_has_jmp_buf__ = 1; - if(setjmp(test_abort_jmp_buf__) != 0) - goto aborted; - } - - test_timer_get_time__(&test_timer_start__); - test->func(); -aborted: - test_abort_has_jmp_buf__ = 0; - test_timer_get_time__(&test_timer_end__); - - if(test_verbose_level__ >= 3) { - test_line_indent__(1); - if(test_current_failures__ == 0) { - test_print_in_color__(TEST_COLOR_GREEN_INTENSIVE__, "SUCCESS: "); - printf("All conditions have passed.\n"); - - if(test_timer__) { - test_line_indent__(1); - printf("Duration: "); - test_timer_print_diff__(); - printf("\n"); - } - } else { - test_print_in_color__(TEST_COLOR_RED_INTENSIVE__, "FAILED: "); - printf("%d condition%s %s failed.\n", - test_current_failures__, - (test_current_failures__ == 1) ? "" : "s", - (test_current_failures__ == 1) ? "has" : "have"); - } - printf("\n"); - } else if(test_verbose_level__ >= 1 && test_current_failures__ == 0) { - test_finish_test_line__(0); - } - - test_case__(NULL); - test_current_unit__ = NULL; - return (test_current_failures__ == 0) ? 0 : -1; - -#ifdef __cplusplus - } catch(std::exception& e) { - const char* what = e.what(); - test_check__(0, NULL, 0, "Threw std::exception"); - if(what != NULL) - test_message__("std::exception::what(): %s", what); - return -1; - } catch(...) { - test_check__(0, NULL, 0, "Threw an exception"); - return -1; - } -#endif -} - -/* Trigger the unit test. If possible (and not suppressed) it starts a child - * process who calls test_do_run__(), otherwise it calls test_do_run__() - * directly. */ -static void -test_run__(const struct test__* test, int index, int master_index) -{ - int failed = 1; - test_timer_type__ start, end; - - test_current_unit__ = test; - test_current_already_logged__ = 0; - test_timer_get_time__(&start); - - if(!test_no_exec__) { - -#if defined(ACUTEST_UNIX__) - - pid_t pid; - int exit_code; - - /* Make sure the child starts with empty I/O buffers. */ - fflush(stdout); - fflush(stderr); - - pid = fork(); - if(pid == (pid_t)-1) { - test_error__("Cannot fork. %s [%d]", strerror(errno), errno); - failed = 1; - } else if(pid == 0) { - /* Child: Do the test. */ - test_worker__ = 1; - failed = (test_do_run__(test, index) != 0); - exit(failed ? 1 : 0); - } else { - /* Parent: Wait until child terminates and analyze its exit code. */ - waitpid(pid, &exit_code, 0); - if(WIFEXITED(exit_code)) { - switch(WEXITSTATUS(exit_code)) { - case 0: failed = 0; break; /* test has passed. */ - case 1: /* noop */ break; /* "normal" failure. */ - default: test_error__("Unexpected exit code [%d]", WEXITSTATUS(exit_code)); - } - } else if(WIFSIGNALED(exit_code)) { - char tmp[32]; - const char* signame; - switch(WTERMSIG(exit_code)) { - case SIGINT: signame = "SIGINT"; break; - case SIGHUP: signame = "SIGHUP"; break; - case SIGQUIT: signame = "SIGQUIT"; break; - case SIGABRT: signame = "SIGABRT"; break; - case SIGKILL: signame = "SIGKILL"; break; - case SIGSEGV: signame = "SIGSEGV"; break; - case SIGILL: signame = "SIGILL"; break; - case SIGTERM: signame = "SIGTERM"; break; - default: sprintf(tmp, "signal %d", WTERMSIG(exit_code)); signame = tmp; break; - } - test_error__("Test interrupted by %s", signame); - } else { - test_error__("Test ended in an unexpected way [%d]", exit_code); - } - } - -#elif defined(ACUTEST_WIN__) - - char buffer[512] = {0}; - STARTUPINFOA startupInfo; - PROCESS_INFORMATION processInfo; - DWORD exitCode; - - /* Windows has no fork(). So we propagate all info into the child - * through a command line arguments. */ - _snprintf(buffer, sizeof(buffer)-1, - "%s --worker=%d %s --no-exec --no-summary %s --verbose=%d --color=%s -- \"%s\"", - test_argv0__, index, test_timer__ ? "--timer" : "", - test_tap__ ? "--tap" : "", test_verbose_level__, - test_colorize__ ? "always" : "never", - test->name); - memset(&startupInfo, 0, sizeof(startupInfo)); - startupInfo.cb = sizeof(STARTUPINFO); - if(CreateProcessA(NULL, buffer, NULL, NULL, FALSE, 0, NULL, NULL, &startupInfo, &processInfo)) { - WaitForSingleObject(processInfo.hProcess, INFINITE); - GetExitCodeProcess(processInfo.hProcess, &exitCode); - CloseHandle(processInfo.hThread); - CloseHandle(processInfo.hProcess); - failed = (exitCode != 0); - } else { - test_error__("Cannot create unit test subprocess [%ld].", GetLastError()); - failed = 1; - } - -#else - - /* A platform where we don't know how to run child process. */ - failed = (test_do_run__(test, index) != 0); - -#endif - - } else { - /* Child processes suppressed through --no-exec. */ - failed = (test_do_run__(test, index) != 0); - } - test_timer_get_time__(&end); - - test_current_unit__ = NULL; - - test_stat_run_units__++; - if(failed) - test_stat_failed_units__++; - - test_set_success__(master_index, !failed); - test_set_duration__(master_index, test_timer_diff__(start, end)); -} - -#if defined(ACUTEST_WIN__) -/* Callback for SEH events. */ -static LONG CALLBACK -test_seh_exception_filter__(EXCEPTION_POINTERS *ptrs) -{ - test_check__(0, NULL, 0, "Unhandled SEH exception"); - test_message__("Exception code: 0x%08lx", ptrs->ExceptionRecord->ExceptionCode); - test_message__("Exception address: 0x%p", ptrs->ExceptionRecord->ExceptionAddress); - - fflush(stdout); - fflush(stderr); - - return EXCEPTION_EXECUTE_HANDLER; -} -#endif - - -#define TEST_CMDLINE_OPTFLAG_OPTIONALARG__ 0x0001 -#define TEST_CMDLINE_OPTFLAG_REQUIREDARG__ 0x0002 - -#define TEST_CMDLINE_OPTID_NONE__ 0 -#define TEST_CMDLINE_OPTID_UNKNOWN__ (-0x7fffffff + 0) -#define TEST_CMDLINE_OPTID_MISSINGARG__ (-0x7fffffff + 1) -#define TEST_CMDLINE_OPTID_BOGUSARG__ (-0x7fffffff + 2) - -typedef struct TEST_CMDLINE_OPTION__ { - char shortname; - const char* longname; - int id; - unsigned flags; -} TEST_CMDLINE_OPTION__; - -static int -test_cmdline_handle_short_opt_group__(const TEST_CMDLINE_OPTION__* options, - const char* arggroup, - int (*callback)(int /*optval*/, const char* /*arg*/)) -{ - const TEST_CMDLINE_OPTION__* opt; - int i; - int ret = 0; - - for(i = 0; arggroup[i] != '\0'; i++) { - for(opt = options; opt->id != 0; opt++) { - if(arggroup[i] == opt->shortname) - break; - } - - if(opt->id != 0 && !(opt->flags & TEST_CMDLINE_OPTFLAG_REQUIREDARG__)) { - ret = callback(opt->id, NULL); - } else { - /* Unknown option. */ - char badoptname[3]; - badoptname[0] = '-'; - badoptname[1] = arggroup[i]; - badoptname[2] = '\0'; - ret = callback((opt->id != 0 ? TEST_CMDLINE_OPTID_MISSINGARG__ : TEST_CMDLINE_OPTID_UNKNOWN__), - badoptname); - } - - if(ret != 0) - break; - } - - return ret; -} - -#define TEST_CMDLINE_AUXBUF_SIZE__ 32 - -static int -test_cmdline_read__(const TEST_CMDLINE_OPTION__* options, int argc, char** argv, - int (*callback)(int /*optval*/, const char* /*arg*/)) -{ - - const TEST_CMDLINE_OPTION__* opt; - char auxbuf[TEST_CMDLINE_AUXBUF_SIZE__+1]; - int after_doubledash = 0; - int i = 1; - int ret = 0; - - auxbuf[TEST_CMDLINE_AUXBUF_SIZE__] = '\0'; - - while(i < argc) { - if(after_doubledash || strcmp(argv[i], "-") == 0) { - /* Non-option argument. */ - ret = callback(TEST_CMDLINE_OPTID_NONE__, argv[i]); - } else if(strcmp(argv[i], "--") == 0) { - /* End of options. All the remaining members are non-option arguments. */ - after_doubledash = 1; - } else if(argv[i][0] != '-') { - /* Non-option argument. */ - ret = callback(TEST_CMDLINE_OPTID_NONE__, argv[i]); - } else { - for(opt = options; opt->id != 0; opt++) { - if(opt->longname != NULL && strncmp(argv[i], "--", 2) == 0) { - size_t len = strlen(opt->longname); - if(strncmp(argv[i]+2, opt->longname, len) == 0) { - /* Regular long option. */ - if(argv[i][2+len] == '\0') { - /* with no argument provided. */ - if(!(opt->flags & TEST_CMDLINE_OPTFLAG_REQUIREDARG__)) - ret = callback(opt->id, NULL); - else - ret = callback(TEST_CMDLINE_OPTID_MISSINGARG__, argv[i]); - break; - } else if(argv[i][2+len] == '=') { - /* with an argument provided. */ - if(opt->flags & (TEST_CMDLINE_OPTFLAG_OPTIONALARG__ | TEST_CMDLINE_OPTFLAG_REQUIREDARG__)) { - ret = callback(opt->id, argv[i]+2+len+1); - } else { - sprintf(auxbuf, "--%s", opt->longname); - ret = callback(TEST_CMDLINE_OPTID_BOGUSARG__, auxbuf); - } - break; - } else { - continue; - } - } - } else if(opt->shortname != '\0' && argv[i][0] == '-') { - if(argv[i][1] == opt->shortname) { - /* Regular short option. */ - if(opt->flags & TEST_CMDLINE_OPTFLAG_REQUIREDARG__) { - if(argv[i][2] != '\0') - ret = callback(opt->id, argv[i]+2); - else if(i+1 < argc) - ret = callback(opt->id, argv[++i]); - else - ret = callback(TEST_CMDLINE_OPTID_MISSINGARG__, argv[i]); - break; - } else { - ret = callback(opt->id, NULL); - - /* There might be more (argument-less) short options - * grouped together. */ - if(ret == 0 && argv[i][2] != '\0') - ret = test_cmdline_handle_short_opt_group__(options, argv[i]+2, callback); - break; - } - } - } - } - - if(opt->id == 0) { /* still not handled? */ - if(argv[i][0] != '-') { - /* Non-option argument. */ - ret = callback(TEST_CMDLINE_OPTID_NONE__, argv[i]); - } else { - /* Unknown option. */ - char* badoptname = argv[i]; - - if(strncmp(badoptname, "--", 2) == 0) { - /* Strip any argument from the long option. */ - char* assignment = strchr(badoptname, '='); - if(assignment != NULL) { - size_t len = assignment - badoptname; - if(len > TEST_CMDLINE_AUXBUF_SIZE__) - len = TEST_CMDLINE_AUXBUF_SIZE__; - strncpy(auxbuf, badoptname, len); - auxbuf[len] = '\0'; - badoptname = auxbuf; - } - } - - ret = callback(TEST_CMDLINE_OPTID_UNKNOWN__, badoptname); - } - } - } - - if(ret != 0) - return ret; - i++; - } - - return ret; -} - -static void -test_help__(void) -{ - printf("Usage: %s [options] [test...]\n", test_argv0__); - printf("\n"); - printf("Run the specified unit tests; or if the option '--skip' is used, run all\n"); - printf("tests in the suite but those listed. By default, if no tests are specified\n"); - printf("on the command line, all unit tests in the suite are run.\n"); - printf("\n"); - printf("Options:\n"); - printf(" -s, --skip Execute all unit tests but the listed ones\n"); - printf(" --exec[=WHEN] If supported, execute unit tests as child processes\n"); - printf(" (WHEN is one of 'auto', 'always', 'never')\n"); -#if defined ACUTEST_WIN__ - printf(" -t, --timer Measure test duration\n"); -#elif defined ACUTEST_HAS_POSIX_TIMER__ - printf(" -t, --timer Measure test duration (real time)\n"); - printf(" --timer=TIMER Measure test duration, using given timer\n"); - printf(" (TIMER is one of 'real', 'cpu')\n"); -#endif - printf(" -E, --no-exec Same as --exec=never\n"); - printf(" --no-summary Suppress printing of test results summary\n"); - printf(" --tap Produce TAP-compliant output\n"); - printf(" (See https://testanything.org/)\n"); - printf(" -x, --xml-output=FILE Enable XUnit output to the given file\n"); - printf(" -l, --list List unit tests in the suite and exit\n"); - printf(" -v, --verbose Make output more verbose\n"); - printf(" --verbose=LEVEL Set verbose level to LEVEL:\n"); - printf(" 0 ... Be silent\n"); - printf(" 1 ... Output one line per test (and summary)\n"); - printf(" 2 ... As 1 and failed conditions (this is default)\n"); - printf(" 3 ... As 1 and all conditions (and extended summary)\n"); - printf(" --color[=WHEN] Enable colorized output\n"); - printf(" (WHEN is one of 'auto', 'always', 'never')\n"); - printf(" --no-color Same as --color=never\n"); - printf(" -h, --help Display this help and exit\n"); - - if(test_list_size__ < 16) { - printf("\n"); - test_list_names__(); - } -} - -static const TEST_CMDLINE_OPTION__ test_cmdline_options__[] = { - { 's', "skip", 's', 0 }, - { 0, "exec", 'e', TEST_CMDLINE_OPTFLAG_OPTIONALARG__ }, - { 'E', "no-exec", 'E', 0 }, -#if defined ACUTEST_WIN__ - { 't', "timer", 't', 0 }, -#elif defined ACUTEST_HAS_POSIX_TIMER__ - { 't', "timer", 't', TEST_CMDLINE_OPTFLAG_OPTIONALARG__ }, -#endif - { 0, "no-summary", 'S', 0 }, - { 0, "tap", 'T', 0 }, - { 'l', "list", 'l', 0 }, - { 'v', "verbose", 'v', TEST_CMDLINE_OPTFLAG_OPTIONALARG__ }, - { 0, "color", 'c', TEST_CMDLINE_OPTFLAG_OPTIONALARG__ }, - { 0, "no-color", 'C', 0 }, - { 'h', "help", 'h', 0 }, - { 0, "worker", 'w', TEST_CMDLINE_OPTFLAG_REQUIREDARG__ }, /* internal */ - { 'x', "xml-output", 'x', TEST_CMDLINE_OPTFLAG_REQUIREDARG__ }, - { 0, NULL, 0, 0 } -}; - -static int -test_cmdline_callback__(int id, const char* arg) -{ - switch(id) { - case 's': - test_skip_mode__ = 1; - break; - - case 'e': - if(arg == NULL || strcmp(arg, "always") == 0) { - test_no_exec__ = 0; - } else if(strcmp(arg, "never") == 0) { - test_no_exec__ = 1; - } else if(strcmp(arg, "auto") == 0) { - /*noop*/ - } else { - fprintf(stderr, "%s: Unrecognized argument '%s' for option --exec.\n", test_argv0__, arg); - fprintf(stderr, "Try '%s --help' for more information.\n", test_argv0__); - exit(2); - } - break; - - case 'E': - test_no_exec__ = 1; - break; - - case 't': -#if defined ACUTEST_WIN__ || defined ACUTEST_HAS_POSIX_TIMER__ - if(arg == NULL || strcmp(arg, "real") == 0) { - test_timer__ = 1; - #ifndef ACUTEST_WIN__ - } else if(strcmp(arg, "cpu") == 0) { - test_timer__ = 2; - #endif - } else { - fprintf(stderr, "%s: Unrecognized argument '%s' for option --timer.\n", test_argv0__, arg); - fprintf(stderr, "Try '%s --help' for more information.\n", test_argv0__); - exit(2); - } -#endif - break; - - case 'S': - test_no_summary__ = 1; - break; - - case 'T': - test_tap__ = 1; - break; - - case 'l': - test_list_names__(); - exit(0); - - case 'v': - test_verbose_level__ = (arg != NULL ? atoi(arg) : test_verbose_level__+1); - break; - - case 'c': - if(arg == NULL || strcmp(arg, "always") == 0) { - test_colorize__ = 1; - } else if(strcmp(arg, "never") == 0) { - test_colorize__ = 0; - } else if(strcmp(arg, "auto") == 0) { - /*noop*/ - } else { - fprintf(stderr, "%s: Unrecognized argument '%s' for option --color.\n", test_argv0__, arg); - fprintf(stderr, "Try '%s --help' for more information.\n", test_argv0__); - exit(2); - } - break; - - case 'C': - test_colorize__ = 0; - break; - - case 'h': - test_help__(); - exit(0); - - case 'w': - test_worker__ = 1; - test_worker_index__ = atoi(arg); - break; - case 'x': - test_xml_output__ = fopen(arg, "w"); - if (!test_xml_output__) { - fprintf(stderr, "Unable to open '%s': %s\n", arg, strerror(errno)); - exit(2); - } - break; - - case 0: - if(test_lookup__(arg) == 0) { - fprintf(stderr, "%s: Unrecognized unit test '%s'\n", test_argv0__, arg); - fprintf(stderr, "Try '%s --list' for list of unit tests.\n", test_argv0__); - exit(2); - } - break; - - case TEST_CMDLINE_OPTID_UNKNOWN__: - fprintf(stderr, "Unrecognized command line option '%s'.\n", arg); - fprintf(stderr, "Try '%s --help' for more information.\n", test_argv0__); - exit(2); - - case TEST_CMDLINE_OPTID_MISSINGARG__: - fprintf(stderr, "The command line option '%s' requires an argument.\n", arg); - fprintf(stderr, "Try '%s --help' for more information.\n", test_argv0__); - exit(2); - - case TEST_CMDLINE_OPTID_BOGUSARG__: - fprintf(stderr, "The command line option '%s' does not expect an argument.\n", arg); - fprintf(stderr, "Try '%s --help' for more information.\n", test_argv0__); - exit(2); - } - - return 0; -} - - -#ifdef ACUTEST_LINUX__ -static int -test_is_tracer_present__(void) -{ - char buf[256+32+1]; - int tracer_present = 0; - int fd; - ssize_t n_read; - - fd = open("/proc/self/status", O_RDONLY); - if(fd == -1) - return 0; - - n_read = read(fd, buf, sizeof(buf)-1); - while(n_read > 0) { - static const char pattern[] = "TracerPid:"; - const char* field; - - buf[n_read] = '\0'; - field = strstr(buf, pattern); - if(field != NULL && field < buf + sizeof(buf) - 32) { - pid_t tracer_pid = (pid_t) atoi(field + sizeof(pattern) - 1); - tracer_present = (tracer_pid != 0); - break; - } - - if(n_read == sizeof(buf)-1) { - memmove(buf, buf + sizeof(buf)-1 - 32, 32); - n_read = read(fd, buf+32, sizeof(buf)-1-32); - if(n_read > 0) - n_read += 32; - } - } - - close(fd); - return tracer_present; -} -#endif - -int -main(int argc, char** argv) -{ - int i; - test_argv0__ = argv[0]; - -#if defined ACUTEST_UNIX__ - test_colorize__ = isatty(STDOUT_FILENO); -#elif defined ACUTEST_WIN__ - #if defined __BORLANDC__ - test_colorize__ = isatty(_fileno(stdout)); - #else - test_colorize__ = _isatty(_fileno(stdout)); - #endif -#else - test_colorize__ = 0; -#endif - - test_timer_init__(); - - /* Count all test units */ - test_list_size__ = 0; - for(i = 0; test_list__[i].func != NULL; i++) - test_list_size__++; - - test_details__ = (struct test_detail__*)calloc(test_list_size__, sizeof(struct test_detail__)); - if(test_details__ == NULL) { - fprintf(stderr, "Out of memory.\n"); - exit(2); - } - - /* Parse options */ - test_cmdline_read__(test_cmdline_options__, argc, argv, test_cmdline_callback__); - -#if defined(ACUTEST_WIN__) - SetUnhandledExceptionFilter(test_seh_exception_filter__); -#endif - - /* By default, we want to run all tests. */ - if(test_count__ == 0) { - for(i = 0; test_list__[i].func != NULL; i++) - test_remember__(i); - } - - /* Guess whether we want to run unit tests as child processes. */ - if(test_no_exec__ < 0) { - test_no_exec__ = 0; - - if(test_count__ <= 1) { - test_no_exec__ = 1; - } else { -#ifdef ACUTEST_WIN__ - if(IsDebuggerPresent()) - test_no_exec__ = 1; -#endif -#ifdef ACUTEST_LINUX__ - if(test_is_tracer_present__()) - test_no_exec__ = 1; -#endif - } - } - - if(test_tap__) { - /* TAP requires we know test result ("ok", "not ok") before we output - * anything about the test, and this gets problematic for larger verbose - * levels. */ - if(test_verbose_level__ > 2) - test_verbose_level__ = 2; - - /* TAP harness should provide some summary. */ - test_no_summary__ = 1; - - if(!test_worker__) - printf("1..%d\n", (int) test_count__); - } - - int index = test_worker_index__; - for(i = 0; test_list__[i].func != NULL; i++) { - int run = (test_details__[i].flags & TEST_FLAG_RUN__); - if (test_skip_mode__) /* Run all tests except those listed. */ - run = !run; - if(run) - test_run__(&test_list__[i], index++, i); - } - - /* Write a summary */ - if(!test_no_summary__ && test_verbose_level__ >= 1) { - if(test_verbose_level__ >= 3) { - test_print_in_color__(TEST_COLOR_DEFAULT_INTENSIVE__, "Summary:\n"); - - printf(" Count of all unit tests: %4d\n", (int) test_list_size__); - printf(" Count of run unit tests: %4d\n", test_stat_run_units__); - printf(" Count of failed unit tests: %4d\n", test_stat_failed_units__); - printf(" Count of skipped unit tests: %4d\n", (int) test_list_size__ - test_stat_run_units__); - } - - if(test_stat_failed_units__ == 0) { - test_print_in_color__(TEST_COLOR_GREEN_INTENSIVE__, "SUCCESS:"); - printf(" All unit tests have passed.\n"); - } else { - test_print_in_color__(TEST_COLOR_RED_INTENSIVE__, "FAILED:"); - printf(" %d of %d unit tests %s failed.\n", - test_stat_failed_units__, test_stat_run_units__, - (test_stat_failed_units__ == 1) ? "has" : "have"); - } - - if(test_verbose_level__ >= 3) - printf("\n"); - } - - if (test_xml_output__) { -#if defined ACUTEST_UNIX__ - char *suite_name = basename(argv[0]); -#elif defined ACUTEST_WIN__ - char suite_name[_MAX_FNAME]; - _splitpath(argv[0], NULL, NULL, suite_name, NULL); -#else - const char *suite_name = argv[0]; -#endif - fprintf(test_xml_output__, "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n"); - fprintf(test_xml_output__, "<testsuite name=\"%s\" tests=\"%d\" errors=\"%d\" failures=\"%d\" skip=\"%d\">\n", - suite_name, (int)test_list_size__, test_stat_failed_units__, test_stat_failed_units__, - (int)test_list_size__ - test_stat_run_units__); - for(i = 0; test_list__[i].func != NULL; i++) { - struct test_detail__ *details = &test_details__[i]; - fprintf(test_xml_output__, " <testcase name=\"%s\" time=\"%.2f\">\n", test_list__[i].name, details->duration); - if (details->flags & TEST_FLAG_FAILURE__) - fprintf(test_xml_output__, " <failure />\n"); - if (!(details->flags & TEST_FLAG_FAILURE__) && !(details->flags & TEST_FLAG_SUCCESS__)) - fprintf(test_xml_output__, " <skipped />\n"); - fprintf(test_xml_output__, " </testcase>\n"); - } - fprintf(test_xml_output__, "</testsuite>\n"); - fclose(test_xml_output__); - } - - free((void*) test_details__); - - return (test_stat_failed_units__ == 0) ? 0 : 1; -} - - -#endif /* #ifndef TEST_NO_MAIN */ - -#ifdef __cplusplus - } /* extern "C" */ -#endif - - -#endif /* #ifndef ACUTEST_H__ */ |
