| /* GLib testing utilities |
| * Copyright (C) 2007 Imendio AB |
| * Authors: Tim Janik, Sven Herzberg |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with this library; if not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include "config.h" |
| |
| #include "gtestutils.h" |
| #include "gfileutils.h" |
| |
| #include <sys/types.h> |
| #ifdef G_OS_UNIX |
| #include <sys/wait.h> |
| #include <sys/time.h> |
| #include <fcntl.h> |
| #include <unistd.h> |
| #include <glib/gstdio.h> |
| #endif |
| #include <string.h> |
| #include <stdlib.h> |
| #include <stdio.h> |
| #ifdef HAVE_SYS_RESOURCE_H |
| #include <sys/resource.h> |
| #endif |
| #ifdef G_OS_WIN32 |
| #include <io.h> |
| #include <windows.h> |
| #endif |
| #include <errno.h> |
| #include <signal.h> |
| #ifdef HAVE_SYS_SELECT_H |
| #include <sys/select.h> |
| #endif /* HAVE_SYS_SELECT_H */ |
| |
| #include "gmain.h" |
| #include "gpattern.h" |
| #include "grand.h" |
| #include "gstrfuncs.h" |
| #include "gtimer.h" |
| #include "gslice.h" |
| #include "gspawn.h" |
| #include "glib-private.h" |
| |
| |
| /** |
| * SECTION:testing |
| * @title: Testing |
| * @short_description: a test framework |
| * @see_also: [gtester][gtester], [gtester-report][gtester-report] |
| * |
| * GLib provides a framework for writing and maintaining unit tests |
| * in parallel to the code they are testing. The API is designed according |
| * to established concepts found in the other test frameworks (JUnit, NUnit, |
| * RUnit), which in turn is based on smalltalk unit testing concepts. |
| * |
| * - Test case: Tests (test methods) are grouped together with their |
| * fixture into test cases. |
| * |
| * - Fixture: A test fixture consists of fixture data and setup and |
| * teardown methods to establish the environment for the test |
| * functions. We use fresh fixtures, i.e. fixtures are newly set |
| * up and torn down around each test invocation to avoid dependencies |
| * between tests. |
| * |
| * - Test suite: Test cases can be grouped into test suites, to allow |
| * subsets of the available tests to be run. Test suites can be |
| * grouped into other test suites as well. |
| * |
| * The API is designed to handle creation and registration of test suites |
| * and test cases implicitly. A simple call like |
| * |[<!-- language="C" --> |
| * g_test_add_func ("/misc/assertions", test_assertions); |
| * ]| |
| * creates a test suite called "misc" with a single test case named |
| * "assertions", which consists of running the test_assertions function. |
| * |
| * In addition to the traditional g_assert_true(), the test framework provides |
| * an extended set of assertions for comparisons: g_assert_cmpfloat(), |
| * g_assert_cmpfloat_with_epsilon(), g_assert_cmpint(), g_assert_cmpuint(), |
| * g_assert_cmphex(), g_assert_cmpstr(), and g_assert_cmpmem(). The |
| * advantage of these variants over plain g_assert_true() is that the assertion |
| * messages can be more elaborate, and include the values of the compared |
| * entities. |
| * |
| * Note that g_assert() should not be used in unit tests, since it is a no-op |
| * when compiling with `G_DISABLE_ASSERT`. Use g_assert() in production code, |
| * and g_assert_true() in unit tests. |
| * |
| * A full example of creating a test suite with two tests using fixtures: |
| * |[<!-- language="C" --> |
| * #include <glib.h> |
| * #include <locale.h> |
| * |
| * typedef struct { |
| * MyObject *obj; |
| * OtherObject *helper; |
| * } MyObjectFixture; |
| * |
| * static void |
| * my_object_fixture_set_up (MyObjectFixture *fixture, |
| * gconstpointer user_data) |
| * { |
| * fixture->obj = my_object_new (); |
| * my_object_set_prop1 (fixture->obj, "some-value"); |
| * my_object_do_some_complex_setup (fixture->obj, user_data); |
| * |
| * fixture->helper = other_object_new (); |
| * } |
| * |
| * static void |
| * my_object_fixture_tear_down (MyObjectFixture *fixture, |
| * gconstpointer user_data) |
| * { |
| * g_clear_object (&fixture->helper); |
| * g_clear_object (&fixture->obj); |
| * } |
| * |
| * static void |
| * test_my_object_test1 (MyObjectFixture *fixture, |
| * gconstpointer user_data) |
| * { |
| * g_assert_cmpstr (my_object_get_property (fixture->obj), ==, "initial-value"); |
| * } |
| * |
| * static void |
| * test_my_object_test2 (MyObjectFixture *fixture, |
| * gconstpointer user_data) |
| * { |
| * my_object_do_some_work_using_helper (fixture->obj, fixture->helper); |
| * g_assert_cmpstr (my_object_get_property (fixture->obj), ==, "updated-value"); |
| * } |
| * |
| * int |
| * main (int argc, char *argv[]) |
| * { |
| * setlocale (LC_ALL, ""); |
| * |
| * g_test_init (&argc, &argv, NULL); |
| * g_test_bug_base ("http://bugzilla.gnome.org/show_bug.cgi?id="); |
| * |
| * // Define the tests. |
| * g_test_add ("/my-object/test1", MyObjectFixture, "some-user-data", |
| * my_object_fixture_set_up, test_my_object_test1, |
| * my_object_fixture_tear_down); |
| * g_test_add ("/my-object/test2", MyObjectFixture, "some-user-data", |
| * my_object_fixture_set_up, test_my_object_test2, |
| * my_object_fixture_tear_down); |
| * |
| * return g_test_run (); |
| * } |
| * ]| |
| * |
| * ### Integrating GTest in your project |
| * |
| * If you are using the [Meson](http://mesonbuild.com) build system, you will |
| * typically use the provided `test()` primitive to call the test binaries, |
| * e.g.: |
| * |
| * |[<!-- language="plain" --> |
| * test( |
| * 'foo', |
| * executable('foo', 'foo.c', dependencies: deps), |
| * env: [ |
| * 'G_TEST_SRCDIR=@0@'.format(meson.current_source_dir()), |
| * 'G_TEST_BUILDDIR=@0@'.format(meson.current_build_dir()), |
| * ], |
| * ) |
| * |
| * test( |
| * 'bar', |
| * executable('bar', 'bar.c', dependencies: deps), |
| * env: [ |
| * 'G_TEST_SRCDIR=@0@'.format(meson.current_source_dir()), |
| * 'G_TEST_BUILDDIR=@0@'.format(meson.current_build_dir()), |
| * ], |
| * ) |
| * ]| |
| * |
| * If you are using Autotools, you're strongly encouraged to use the Automake |
| * [TAP](https://testanything.org/) harness; GLib provides template files for |
| * easily integrating with it: |
| * |
| * - [glib-tap.mk](https://git.gnome.org/browse/glib/tree/glib-tap.mk) |
| * - [tap-test](https://git.gnome.org/browse/glib/tree/tap-test) |
| * - [tap-driver.sh](https://git.gnome.org/browse/glib/tree/tap-driver.sh) |
| * |
| * You can copy these files in your own project's root directory, and then |
| * set up your `Makefile.am` file to reference them, for instance: |
| * |
| * |[<!-- language="plain" --> |
| * include $(top_srcdir)/glib-tap.mk |
| * |
| * # test binaries |
| * test_programs = \ |
| * foo \ |
| * bar |
| * |
| * # data distributed in the tarball |
| * dist_test_data = \ |
| * foo.data.txt \ |
| * bar.data.txt |
| * |
| * # data not distributed in the tarball |
| * test_data = \ |
| * blah.data.txt |
| * ]| |
| * |
| * Make sure to distribute the TAP files, using something like the following |
| * in your top-level `Makefile.am`: |
| * |
| * |[<!-- language="plain" --> |
| * EXTRA_DIST += \ |
| * tap-driver.sh \ |
| * tap-test |
| * ]| |
| * |
| * `glib-tap.mk` will be distributed implicitly due to being included in a |
| * `Makefile.am`. All three files should be added to version control. |
| * |
| * If you don't have access to the Autotools TAP harness, you can use the |
| * [gtester][gtester] and [gtester-report][gtester-report] tools, and use |
| * the [glib.mk](https://git.gnome.org/browse/glib/tree/glib.mk) Automake |
| * template provided by GLib. |
| */ |
| |
| /** |
| * g_test_initialized: |
| * |
| * Returns %TRUE if g_test_init() has been called. |
| * |
| * Returns: %TRUE if g_test_init() has been called. |
| * |
| * Since: 2.36 |
| */ |
| |
| /** |
| * g_test_quick: |
| * |
| * Returns %TRUE if tests are run in quick mode. |
| * Exactly one of g_test_quick() and g_test_slow() is active in any run; |
| * there is no "medium speed". |
| * |
| * By default, tests are run in quick mode. In tests that use |
| * g_test_init(), the options `-m quick`, `-m slow` and `-m thorough` |
| * can be used to change this. |
| * |
| * Returns: %TRUE if in quick mode |
| */ |
| |
| /** |
| * g_test_slow: |
| * |
| * Returns %TRUE if tests are run in slow mode. |
| * Exactly one of g_test_quick() and g_test_slow() is active in any run; |
| * there is no "medium speed". |
| * |
| * By default, tests are run in quick mode. In tests that use |
| * g_test_init(), the options `-m quick`, `-m slow` and `-m thorough` |
| * can be used to change this. |
| * |
| * Returns: the opposite of g_test_quick() |
| */ |
| |
| /** |
| * g_test_thorough: |
| * |
| * Returns %TRUE if tests are run in thorough mode, equivalent to |
| * g_test_slow(). |
| * |
| * By default, tests are run in quick mode. In tests that use |
| * g_test_init(), the options `-m quick`, `-m slow` and `-m thorough` |
| * can be used to change this. |
| * |
| * Returns: the same thing as g_test_slow() |
| */ |
| |
| /** |
| * g_test_perf: |
| * |
| * Returns %TRUE if tests are run in performance mode. |
| * |
| * By default, tests are run in quick mode. In tests that use |
| * g_test_init(), the option `-m perf` enables performance tests, while |
| * `-m quick` disables them. |
| * |
| * Returns: %TRUE if in performance mode |
| */ |
| |
| /** |
| * g_test_undefined: |
| * |
| * Returns %TRUE if tests may provoke assertions and other formally-undefined |
| * behaviour, to verify that appropriate warnings are given. It might, in some |
| * cases, be useful to turn this off with if running tests under valgrind; |
| * in tests that use g_test_init(), the option `-m no-undefined` disables |
| * those tests, while `-m undefined` explicitly enables them (the default |
| * behaviour). |
| * |
| * Returns: %TRUE if tests may provoke programming errors |
| */ |
| |
| /** |
| * g_test_verbose: |
| * |
| * Returns %TRUE if tests are run in verbose mode. |
| * In tests that use g_test_init(), the option `--verbose` enables this, |
| * while `-q` or `--quiet` disables it. |
| * The default is neither g_test_verbose() nor g_test_quiet(). |
| * |
| * Returns: %TRUE if in verbose mode |
| */ |
| |
| /** |
| * g_test_quiet: |
| * |
| * Returns %TRUE if tests are run in quiet mode. |
| * In tests that use g_test_init(), the option `-q` or `--quiet` enables |
| * this, while `--verbose` disables it. |
| * The default is neither g_test_verbose() nor g_test_quiet(). |
| * |
| * Returns: %TRUE if in quiet mode |
| */ |
| |
| /** |
| * g_test_queue_unref: |
| * @gobject: the object to unref |
| * |
| * Enqueue an object to be released with g_object_unref() during |
| * the next teardown phase. This is equivalent to calling |
| * g_test_queue_destroy() with a destroy callback of g_object_unref(). |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * GTestTrapFlags: |
| * @G_TEST_TRAP_SILENCE_STDOUT: Redirect stdout of the test child to |
| * `/dev/null` so it cannot be observed on the console during test |
| * runs. The actual output is still captured though to allow later |
| * tests with g_test_trap_assert_stdout(). |
| * @G_TEST_TRAP_SILENCE_STDERR: Redirect stderr of the test child to |
| * `/dev/null` so it cannot be observed on the console during test |
| * runs. The actual output is still captured though to allow later |
| * tests with g_test_trap_assert_stderr(). |
| * @G_TEST_TRAP_INHERIT_STDIN: If this flag is given, stdin of the |
| * child process is shared with stdin of its parent process. |
| * It is redirected to `/dev/null` otherwise. |
| * |
| * Test traps are guards around forked tests. |
| * These flags determine what traps to set. |
| * |
| * Deprecated: #GTestTrapFlags is used only with g_test_trap_fork(), |
| * which is deprecated. g_test_trap_subprocess() uses |
| * #GTestSubprocessFlags. |
| */ |
| |
| /** |
| * GTestSubprocessFlags: |
| * @G_TEST_SUBPROCESS_INHERIT_STDIN: If this flag is given, the child |
| * process will inherit the parent's stdin. Otherwise, the child's |
| * stdin is redirected to `/dev/null`. |
| * @G_TEST_SUBPROCESS_INHERIT_STDOUT: If this flag is given, the child |
| * process will inherit the parent's stdout. Otherwise, the child's |
| * stdout will not be visible, but it will be captured to allow |
| * later tests with g_test_trap_assert_stdout(). |
| * @G_TEST_SUBPROCESS_INHERIT_STDERR: If this flag is given, the child |
| * process will inherit the parent's stderr. Otherwise, the child's |
| * stderr will not be visible, but it will be captured to allow |
| * later tests with g_test_trap_assert_stderr(). |
| * |
| * Flags to pass to g_test_trap_subprocess() to control input and output. |
| * |
| * Note that in contrast with g_test_trap_fork(), the default is to |
| * not show stdout and stderr. |
| */ |
| |
| /** |
| * g_test_trap_assert_passed: |
| * |
| * Assert that the last test subprocess passed. |
| * See g_test_trap_subprocess(). |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_test_trap_assert_failed: |
| * |
| * Assert that the last test subprocess failed. |
| * See g_test_trap_subprocess(). |
| * |
| * This is sometimes used to test situations that are formally considered to |
| * be undefined behaviour, like inputs that fail a g_return_if_fail() |
| * check. In these situations you should skip the entire test, including the |
| * call to g_test_trap_subprocess(), unless g_test_undefined() returns %TRUE |
| * to indicate that undefined behaviour may be tested. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_test_trap_assert_stdout: |
| * @soutpattern: a glob-style [pattern][glib-Glob-style-pattern-matching] |
| * |
| * Assert that the stdout output of the last test subprocess matches |
| * @soutpattern. See g_test_trap_subprocess(). |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_test_trap_assert_stdout_unmatched: |
| * @soutpattern: a glob-style [pattern][glib-Glob-style-pattern-matching] |
| * |
| * Assert that the stdout output of the last test subprocess |
| * does not match @soutpattern. See g_test_trap_subprocess(). |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_test_trap_assert_stderr: |
| * @serrpattern: a glob-style [pattern][glib-Glob-style-pattern-matching] |
| * |
| * Assert that the stderr output of the last test subprocess |
| * matches @serrpattern. See g_test_trap_subprocess(). |
| * |
| * This is sometimes used to test situations that are formally |
| * considered to be undefined behaviour, like code that hits a |
| * g_assert() or g_error(). In these situations you should skip the |
| * entire test, including the call to g_test_trap_subprocess(), unless |
| * g_test_undefined() returns %TRUE to indicate that undefined |
| * behaviour may be tested. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_test_trap_assert_stderr_unmatched: |
| * @serrpattern: a glob-style [pattern][glib-Glob-style-pattern-matching] |
| * |
| * Assert that the stderr output of the last test subprocess |
| * does not match @serrpattern. See g_test_trap_subprocess(). |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_test_rand_bit: |
| * |
| * Get a reproducible random bit (0 or 1), see g_test_rand_int() |
| * for details on test case random numbers. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_assert: |
| * @expr: the expression to check |
| * |
| * Debugging macro to terminate the application if the assertion |
| * fails. If the assertion fails (i.e. the expression is not true), |
| * an error message is logged and the application is terminated. |
| * |
| * The macro can be turned off in final releases of code by defining |
| * `G_DISABLE_ASSERT` when compiling the application, so code must |
| * not depend on any side effects from @expr. Similarly, it must not be used |
| * in unit tests, otherwise the unit tests will be ineffective if compiled with |
| * `G_DISABLE_ASSERT`. Use g_assert_true() and related macros in unit tests |
| * instead. |
| */ |
| |
| /** |
| * g_assert_not_reached: |
| * |
| * Debugging macro to terminate the application if it is ever |
| * reached. If it is reached, an error message is logged and the |
| * application is terminated. |
| * |
| * The macro can be turned off in final releases of code by defining |
| * `G_DISABLE_ASSERT` when compiling the application. Hence, it should not be |
| * used in unit tests, where assertions should always be effective. |
| */ |
| |
| /** |
| * g_assert_true: |
| * @expr: the expression to check |
| * |
| * Debugging macro to check that an expression is true. |
| * |
| * If the assertion fails (i.e. the expression is not true), |
| * an error message is logged and the application is either |
| * terminated or the testcase marked as failed. |
| * |
| * Note that unlike g_assert(), this macro is unaffected by whether |
| * `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and, |
| * conversely, g_assert() should not be used in tests. |
| * |
| * See g_test_set_nonfatal_assertions(). |
| * |
| * Since: 2.38 |
| */ |
| |
| /** |
| * g_assert_false: |
| * @expr: the expression to check |
| * |
| * Debugging macro to check an expression is false. |
| * |
| * If the assertion fails (i.e. the expression is not false), |
| * an error message is logged and the application is either |
| * terminated or the testcase marked as failed. |
| * |
| * Note that unlike g_assert(), this macro is unaffected by whether |
| * `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and, |
| * conversely, g_assert() should not be used in tests. |
| * |
| * See g_test_set_nonfatal_assertions(). |
| * |
| * Since: 2.38 |
| */ |
| |
| /** |
| * g_assert_null: |
| * @expr: the expression to check |
| * |
| * Debugging macro to check an expression is %NULL. |
| * |
| * If the assertion fails (i.e. the expression is not %NULL), |
| * an error message is logged and the application is either |
| * terminated or the testcase marked as failed. |
| * |
| * Note that unlike g_assert(), this macro is unaffected by whether |
| * `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and, |
| * conversely, g_assert() should not be used in tests. |
| * |
| * See g_test_set_nonfatal_assertions(). |
| * |
| * Since: 2.38 |
| */ |
| |
| /** |
| * g_assert_nonnull: |
| * @expr: the expression to check |
| * |
| * Debugging macro to check an expression is not %NULL. |
| * |
| * If the assertion fails (i.e. the expression is %NULL), |
| * an error message is logged and the application is either |
| * terminated or the testcase marked as failed. |
| * |
| * Note that unlike g_assert(), this macro is unaffected by whether |
| * `G_DISABLE_ASSERT` is defined. Hence it should only be used in tests and, |
| * conversely, g_assert() should not be used in tests. |
| * |
| * See g_test_set_nonfatal_assertions(). |
| * |
| * Since: 2.40 |
| */ |
| |
| /** |
| * g_assert_cmpstr: |
| * @s1: a string (may be %NULL) |
| * @cmp: The comparison operator to use. |
| * One of ==, !=, <, >, <=, >=. |
| * @s2: another string (may be %NULL) |
| * |
| * Debugging macro to compare two strings. If the comparison fails, |
| * an error message is logged and the application is either terminated |
| * or the testcase marked as failed. |
| * The strings are compared using g_strcmp0(). |
| * |
| * The effect of `g_assert_cmpstr (s1, op, s2)` is |
| * the same as `g_assert_true (g_strcmp0 (s1, s2) op 0)`. |
| * The advantage of this macro is that it can produce a message that |
| * includes the actual values of @s1 and @s2. |
| * |
| * |[<!-- language="C" --> |
| * g_assert_cmpstr (mystring, ==, "fubar"); |
| * ]| |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_assert_cmpint: |
| * @n1: an integer |
| * @cmp: The comparison operator to use. |
| * One of ==, !=, <, >, <=, >=. |
| * @n2: another integer |
| * |
| * Debugging macro to compare two integers. |
| * |
| * The effect of `g_assert_cmpint (n1, op, n2)` is |
| * the same as `g_assert_true (n1 op n2)`. The advantage |
| * of this macro is that it can produce a message that includes the |
| * actual values of @n1 and @n2. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_assert_cmpuint: |
| * @n1: an unsigned integer |
| * @cmp: The comparison operator to use. |
| * One of ==, !=, <, >, <=, >=. |
| * @n2: another unsigned integer |
| * |
| * Debugging macro to compare two unsigned integers. |
| * |
| * The effect of `g_assert_cmpuint (n1, op, n2)` is |
| * the same as `g_assert_true (n1 op n2)`. The advantage |
| * of this macro is that it can produce a message that includes the |
| * actual values of @n1 and @n2. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_assert_cmphex: |
| * @n1: an unsigned integer |
| * @cmp: The comparison operator to use. |
| * One of ==, !=, <, >, <=, >=. |
| * @n2: another unsigned integer |
| * |
| * Debugging macro to compare to unsigned integers. |
| * |
| * This is a variant of g_assert_cmpuint() that displays the numbers |
| * in hexadecimal notation in the message. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_assert_cmpfloat: |
| * @n1: an floating point number |
| * @cmp: The comparison operator to use. |
| * One of ==, !=, <, >, <=, >=. |
| * @n2: another floating point number |
| * |
| * Debugging macro to compare two floating point numbers. |
| * |
| * The effect of `g_assert_cmpfloat (n1, op, n2)` is |
| * the same as `g_assert_true (n1 op n2)`. The advantage |
| * of this macro is that it can produce a message that includes the |
| * actual values of @n1 and @n2. |
| * |
| * Since: 2.16 |
| */ |
| |
| /** |
| * g_assert_cmpfloat_with_epsilon: |
| * @n1: an floating point number |
| * @n2: another floating point number |
| * @epsilon: a numeric value that expresses the expected tolerance |
| * between @n1 and @n2 |
| * |
| * Debugging macro to compare two floating point numbers within an epsilon. |
| * |
| * The effect of `g_assert_cmpfloat_with_epsilon (n1, n2, epsilon)` is |
| * the same as `g_assert_true (abs (n1 - n2) < epsilon)`. The advantage |
| * of this macro is that it can produce a message that includes the |
| * actual values of @n1 and @n2. |
| * |
| * Since: 2.58 |
| */ |
| |
| /** |
| * g_assert_cmpmem: |
| * @m1: pointer to a buffer |
| * @l1: length of @m1 |
| * @m2: pointer to another buffer |
| * @l2: length of @m2 |
| * |
| * Debugging macro to compare memory regions. If the comparison fails, |
| * an error message is logged and the application is either terminated |
| * or the testcase marked as failed. |
| * |
| * The effect of `g_assert_cmpmem (m1, l1, m2, l2)` is |
| * the same as `g_assert_true (l1 == l2 && memcmp (m1, m2, l1) == 0)`. |
| * The advantage of this macro is that it can produce a message that |
| * includes the actual values of @l1 and @l2. |
| * |
| * |[<!-- language="C" --> |
| * g_assert_cmpmem (buf->data, buf->len, expected, sizeof (expected)); |
| * ]| |
| * |
| * Since: 2.46 |
| */ |
| |
| /** |
| * g_assert_no_error: |
| * @err: a #GError, possibly %NULL |
| * |
| * Debugging macro to check that a #GError is not set. |
| * |
| * The effect of `g_assert_no_error (err)` is |
| * the same as `g_assert_true (err == NULL)`. The advantage |
| * of this macro is that it can produce a message that includes |
| * the error message and code. |
| * |
| * Since: 2.20 |
| */ |
| |
| /** |
| * g_assert_error: |
| * @err: a #GError, possibly %NULL |
| * @dom: the expected error domain (a #GQuark) |
| * @c: the expected error code |
| * |
| * Debugging macro to check that a method has returned |
| * the correct #GError. |
| * |
| * The effect of `g_assert_error (err, dom, c)` is |
| * the same as `g_assert_true (err != NULL && err->domain |
| * == dom && err->code == c)`. The advantage of this |
| * macro is that it can produce a message that includes the incorrect |
| * error message and code. |
| * |
| * This can only be used to test for a specific error. If you want to |
| * test that @err is set, but don't care what it's set to, just use |
| * `g_assert (err != NULL)` |
| * |
| * Since: 2.20 |
| */ |
| |
| /** |
| * GTestCase: |
| * |
| * An opaque structure representing a test case. |
| */ |
| |
| /** |
| * GTestSuite: |
| * |
| * An opaque structure representing a test suite. |
| */ |
| |
| |
| /* Global variable for storing assertion messages; this is the counterpart to |
| * glibc's (private) __abort_msg variable, and allows developers and crash |
| * analysis systems like Apport and ABRT to fish out assertion messages from |
| * core dumps, instead of having to catch them on screen output. |
| */ |
| GLIB_VAR char *__glib_assert_msg; |
| char *__glib_assert_msg = NULL; |
| |
| /* --- constants --- */ |
| #define G_TEST_STATUS_TIMED_OUT 1024 |
| |
| /* --- structures --- */ |
| struct GTestCase |
| { |
| gchar *name; |
| guint fixture_size; |
| void (*fixture_setup) (void*, gconstpointer); |
| void (*fixture_test) (void*, gconstpointer); |
| void (*fixture_teardown) (void*, gconstpointer); |
| gpointer test_data; |
| }; |
| struct GTestSuite |
| { |
| gchar *name; |
| GSList *suites; |
| GSList *cases; |
| }; |
| typedef struct DestroyEntry DestroyEntry; |
| struct DestroyEntry |
| { |
| DestroyEntry *next; |
| GDestroyNotify destroy_func; |
| gpointer destroy_data; |
| }; |
| |
| /* --- prototypes --- */ |
| static void test_run_seed (const gchar *rseed); |
| static void test_trap_clear (void); |
| static guint8* g_test_log_dump (GTestLogMsg *msg, |
| guint *len); |
| static void gtest_default_log_handler (const gchar *log_domain, |
| GLogLevelFlags log_level, |
| const gchar *message, |
| gpointer unused_data); |
| |
| |
| static const char * const g_test_result_names[] = { |
| "OK", |
| "SKIP", |
| "FAIL", |
| "TODO" |
| }; |
| |
| /* --- variables --- */ |
| static int test_log_fd = -1; |
| static gboolean test_mode_fatal = TRUE; |
| static gboolean g_test_run_once = TRUE; |
| static gboolean test_run_list = FALSE; |
| static gchar *test_run_seedstr = NULL; |
| static GRand *test_run_rand = NULL; |
| static gchar *test_run_name = ""; |
| static GSList **test_filename_free_list; |
| static guint test_run_forks = 0; |
| static guint test_run_count = 0; |
| static guint test_count = 0; |
| static guint test_skipped_count = 0; |
| static GTestResult test_run_success = G_TEST_RUN_FAILURE; |
| static gchar *test_run_msg = NULL; |
| static guint test_startup_skip_count = 0; |
| static GTimer *test_user_timer = NULL; |
| static double test_user_stamp = 0; |
| static GSList *test_paths = NULL; |
| static GSList *test_paths_skipped = NULL; |
| static GTestSuite *test_suite_root = NULL; |
| static int test_trap_last_status = 0; /* unmodified platform-specific status */ |
| static GPid test_trap_last_pid = 0; |
| static char *test_trap_last_subprocess = NULL; |
| static char *test_trap_last_stdout = NULL; |
| static char *test_trap_last_stderr = NULL; |
| static char *test_uri_base = NULL; |
| static gboolean test_debug_log = FALSE; |
| static gboolean test_tap_log = FALSE; |
| static gboolean test_nonfatal_assertions = FALSE; |
| static DestroyEntry *test_destroy_queue = NULL; |
| static char *test_argv0 = NULL; |
| static char *test_argv0_dirname; |
| static const char *test_disted_files_dir; |
| static const char *test_built_files_dir; |
| static char *test_initial_cwd = NULL; |
| static gboolean test_in_forked_child = FALSE; |
| static gboolean test_in_subprocess = FALSE; |
| static GTestConfig mutable_test_config_vars = { |
| FALSE, /* test_initialized */ |
| TRUE, /* test_quick */ |
| FALSE, /* test_perf */ |
| FALSE, /* test_verbose */ |
| FALSE, /* test_quiet */ |
| TRUE, /* test_undefined */ |
| }; |
| const GTestConfig * const g_test_config_vars = &mutable_test_config_vars; |
| static gboolean no_g_set_prgname = FALSE; |
| |
| /* --- functions --- */ |
| const char* |
| g_test_log_type_name (GTestLogType log_type) |
| { |
| switch (log_type) |
| { |
| case G_TEST_LOG_NONE: return "none"; |
| case G_TEST_LOG_ERROR: return "error"; |
| case G_TEST_LOG_START_BINARY: return "binary"; |
| case G_TEST_LOG_LIST_CASE: return "list"; |
| case G_TEST_LOG_SKIP_CASE: return "skip"; |
| case G_TEST_LOG_START_CASE: return "start"; |
| case G_TEST_LOG_STOP_CASE: return "stop"; |
| case G_TEST_LOG_MIN_RESULT: return "minperf"; |
| case G_TEST_LOG_MAX_RESULT: return "maxperf"; |
| case G_TEST_LOG_MESSAGE: return "message"; |
| case G_TEST_LOG_START_SUITE: return "start suite"; |
| case G_TEST_LOG_STOP_SUITE: return "stop suite"; |
| } |
| return "???"; |
| } |
| |
| static void |
| g_test_log_send (guint n_bytes, |
| const guint8 *buffer) |
| { |
| if (test_log_fd >= 0) |
| { |
| int r; |
| do |
| r = write (test_log_fd, buffer, n_bytes); |
| while (r < 0 && errno == EINTR); |
| } |
| if (test_debug_log) |
| { |
| GTestLogBuffer *lbuffer = g_test_log_buffer_new (); |
| GTestLogMsg *msg; |
| guint ui; |
| g_test_log_buffer_push (lbuffer, n_bytes, buffer); |
| msg = g_test_log_buffer_pop (lbuffer); |
| g_warn_if_fail (msg != NULL); |
| g_warn_if_fail (lbuffer->data->len == 0); |
| g_test_log_buffer_free (lbuffer); |
| /* print message */ |
| g_printerr ("{*LOG(%s)", g_test_log_type_name (msg->log_type)); |
| for (ui = 0; ui < msg->n_strings; ui++) |
| g_printerr (":{%s}", msg->strings[ui]); |
| if (msg->n_nums) |
| { |
| g_printerr (":("); |
| for (ui = 0; ui < msg->n_nums; ui++) |
| { |
| if ((long double) (long) msg->nums[ui] == msg->nums[ui]) |
| g_printerr ("%s%ld", ui ? ";" : "", (long) msg->nums[ui]); |
| else |
| g_printerr ("%s%.16g", ui ? ";" : "", (double) msg->nums[ui]); |
| } |
| g_printerr (")"); |
| } |
| g_printerr (":LOG*}\n"); |
| g_test_log_msg_free (msg); |
| } |
| } |
| |
| static void |
| g_test_log (GTestLogType lbit, |
| const gchar *string1, |
| const gchar *string2, |
| guint n_args, |
| long double *largs) |
| { |
| GTestResult result; |
| gboolean fail; |
| GTestLogMsg msg; |
| gchar *astrings[3] = { NULL, NULL, NULL }; |
| guint8 *dbuffer; |
| guint32 dbufferlen; |
| |
| switch (lbit) |
| { |
| case G_TEST_LOG_START_BINARY: |
| if (test_tap_log) |
| g_print ("# random seed: %s\n", string2); |
| else if (g_test_verbose ()) |
| g_print ("GTest: random seed: %s\n", string2); |
| break; |
| case G_TEST_LOG_START_SUITE: |
| if (test_tap_log) |
| { |
| if (string1[0] != 0) |
| g_print ("# Start of %s tests\n", string1); |
| else |
| g_print ("1..%d\n", test_count); |
| } |
| break; |
| case G_TEST_LOG_STOP_SUITE: |
| if (test_tap_log) |
| { |
| if (string1[0] != 0) |
| g_print ("# End of %s tests\n", string1); |
| } |
| break; |
| case G_TEST_LOG_STOP_CASE: |
| result = largs[0]; |
| fail = result == G_TEST_RUN_FAILURE; |
| if (test_tap_log) |
| { |
| g_print ("%s %d %s", fail ? "not ok" : "ok", test_run_count, string1); |
| if (result == G_TEST_RUN_INCOMPLETE) |
| g_print (" # TODO %s\n", string2 ? string2 : ""); |
| else if (result == G_TEST_RUN_SKIPPED) |
| g_print (" # SKIP %s\n", string2 ? string2 : ""); |
| else |
| g_print ("\n"); |
| } |
| else if (g_test_verbose ()) |
| g_print ("GTest: result: %s\n", g_test_result_names[result]); |
| else if (!g_test_quiet ()) |
| g_print ("%s\n", g_test_result_names[result]); |
| if (fail && test_mode_fatal) |
| { |
| if (test_tap_log) |
| g_print ("Bail out!\n"); |
| g_abort (); |
| } |
| if (result == G_TEST_RUN_SKIPPED) |
| test_skipped_count++; |
| break; |
| case G_TEST_LOG_MIN_RESULT: |
| if (test_tap_log) |
| g_print ("# min perf: %s\n", string1); |
| else if (g_test_verbose ()) |
| g_print ("(MINPERF:%s)\n", string1); |
| break; |
| case G_TEST_LOG_MAX_RESULT: |
| if (test_tap_log) |
| g_print ("# max perf: %s\n", string1); |
| else if (g_test_verbose ()) |
| g_print ("(MAXPERF:%s)\n", string1); |
| break; |
| case G_TEST_LOG_MESSAGE: |
| if (test_tap_log) |
| g_print ("# %s\n", string1); |
| else if (g_test_verbose ()) |
| g_print ("(MSG: %s)\n", string1); |
| break; |
| case G_TEST_LOG_ERROR: |
| if (test_tap_log) |
| g_print ("Bail out! %s\n", string1); |
| else if (g_test_verbose ()) |
| g_print ("(ERROR: %s)\n", string1); |
| break; |
| default: ; |
| } |
| |
| msg.log_type = lbit; |
| msg.n_strings = (string1 != NULL) + (string1 && string2); |
| msg.strings = astrings; |
| astrings[0] = (gchar*) string1; |
| astrings[1] = astrings[0] ? (gchar*) string2 : NULL; |
| msg.n_nums = n_args; |
| msg.nums = largs; |
| dbuffer = g_test_log_dump (&msg, &dbufferlen); |
| g_test_log_send (dbufferlen, dbuffer); |
| g_free (dbuffer); |
| |
| switch (lbit) |
| { |
| case G_TEST_LOG_START_CASE: |
| if (test_tap_log) |
| ; |
| else if (g_test_verbose ()) |
| g_print ("GTest: run: %s\n", string1); |
| else if (!g_test_quiet ()) |
| g_print ("%s: ", string1); |
| break; |
| default: ; |
| } |
| } |
| |
| /* We intentionally parse the command line without GOptionContext |
| * because otherwise you would never be able to test it. |
| */ |
| static void |
| parse_args (gint *argc_p, |
| gchar ***argv_p) |
| { |
| guint argc = *argc_p; |
| gchar **argv = *argv_p; |
| guint i, e; |
| |
| test_argv0 = argv[0]; |
| test_initial_cwd = g_get_current_dir (); |
| |
| /* parse known args */ |
| for (i = 1; i < argc; i++) |
| { |
| if (strcmp (argv[i], "--g-fatal-warnings") == 0) |
| { |
| GLogLevelFlags fatal_mask = (GLogLevelFlags) g_log_set_always_fatal ((GLogLevelFlags) G_LOG_FATAL_MASK); |
| fatal_mask = (GLogLevelFlags) (fatal_mask | G_LOG_LEVEL_WARNING | G_LOG_LEVEL_CRITICAL); |
| g_log_set_always_fatal (fatal_mask); |
| argv[i] = NULL; |
| } |
| else if (strcmp (argv[i], "--keep-going") == 0 || |
| strcmp (argv[i], "-k") == 0) |
| { |
| test_mode_fatal = FALSE; |
| argv[i] = NULL; |
| } |
| else if (strcmp (argv[i], "--debug-log") == 0) |
| { |
| test_debug_log = TRUE; |
| argv[i] = NULL; |
| } |
| else if (strcmp (argv[i], "--tap") == 0) |
| { |
| test_tap_log = TRUE; |
| argv[i] = NULL; |
| } |
| else if (strcmp ("--GTestLogFD", argv[i]) == 0 || strncmp ("--GTestLogFD=", argv[i], 13) == 0) |
| { |
| gchar *equal = argv[i] + 12; |
| if (*equal == '=') |
| test_log_fd = g_ascii_strtoull (equal + 1, NULL, 0); |
| else if (i + 1 < argc) |
| { |
| argv[i++] = NULL; |
| test_log_fd = g_ascii_strtoull (argv[i], NULL, 0); |
| } |
| argv[i] = NULL; |
| } |
| else if (strcmp ("--GTestSkipCount", argv[i]) == 0 || strncmp ("--GTestSkipCount=", argv[i], 17) == 0) |
| { |
| gchar *equal = argv[i] + 16; |
| if (*equal == '=') |
| test_startup_skip_count = g_ascii_strtoull (equal + 1, NULL, 0); |
| else if (i + 1 < argc) |
| { |
| argv[i++] = NULL; |
| test_startup_skip_count = g_ascii_strtoull (argv[i], NULL, 0); |
| } |
| argv[i] = NULL; |
| } |
| else if (strcmp ("--GTestSubprocess", argv[i]) == 0) |
| { |
| test_in_subprocess = TRUE; |
| /* We typically expect these child processes to crash, and some |
| * tests spawn a *lot* of them. Avoid spamming system crash |
| * collection programs such as systemd-coredump and abrt. |
| */ |
| #ifdef HAVE_SYS_RESOURCE_H |
| { |
| struct rlimit limit = { 0, 0 }; |
| (void) setrlimit (RLIMIT_CORE, &limit); |
| } |
| #endif |
| argv[i] = NULL; |
| } |
| else if (strcmp ("-p", argv[i]) == 0 || strncmp ("-p=", argv[i], 3) == 0) |
| { |
| gchar *equal = argv[i] + 2; |
| if (*equal == '=') |
| test_paths = g_slist_prepend (test_paths, equal + 1); |
| else if (i + 1 < argc) |
| { |
| argv[i++] = NULL; |
| test_paths = g_slist_prepend (test_paths, argv[i]); |
| } |
| argv[i] = NULL; |
| } |
| else if (strcmp ("-s", argv[i]) == 0 || strncmp ("-s=", argv[i], 3) == 0) |
| { |
| gchar *equal = argv[i] + 2; |
| if (*equal == '=') |
| test_paths_skipped = g_slist_prepend (test_paths_skipped, equal + 1); |
| else if (i + 1 < argc) |
| { |
| argv[i++] = NULL; |
| test_paths_skipped = g_slist_prepend (test_paths_skipped, argv[i]); |
| } |
| argv[i] = NULL; |
| } |
| else if (strcmp ("-m", argv[i]) == 0 || strncmp ("-m=", argv[i], 3) == 0) |
| { |
| gchar *equal = argv[i] + 2; |
| const gchar *mode = ""; |
| if (*equal == '=') |
| mode = equal + 1; |
| else if (i + 1 < argc) |
| { |
| argv[i++] = NULL; |
| mode = argv[i]; |
| } |
| if (strcmp (mode, "perf") == 0) |
| mutable_test_config_vars.test_perf = TRUE; |
| else if (strcmp (mode, "slow") == 0) |
| mutable_test_config_vars.test_quick = FALSE; |
| else if (strcmp (mode, "thorough") == 0) |
| mutable_test_config_vars.test_quick = FALSE; |
| else if (strcmp (mode, "quick") == 0) |
| { |
| mutable_test_config_vars.test_quick = TRUE; |
| mutable_test_config_vars.test_perf = FALSE; |
| } |
| else if (strcmp (mode, "undefined") == 0) |
| mutable_test_config_vars.test_undefined = TRUE; |
| else if (strcmp (mode, "no-undefined") == 0) |
| mutable_test_config_vars.test_undefined = FALSE; |
| else |
| g_error ("unknown test mode: -m %s", mode); |
| argv[i] = NULL; |
| } |
| else if (strcmp ("-q", argv[i]) == 0 || strcmp ("--quiet", argv[i]) == 0) |
| { |
| mutable_test_config_vars.test_quiet = TRUE; |
| mutable_test_config_vars.test_verbose = FALSE; |
| argv[i] = NULL; |
| } |
| else if (strcmp ("--verbose", argv[i]) == 0) |
| { |
| mutable_test_config_vars.test_quiet = FALSE; |
| mutable_test_config_vars.test_verbose = TRUE; |
| argv[i] = NULL; |
| } |
| else if (strcmp ("-l", argv[i]) == 0) |
| { |
| test_run_list = TRUE; |
| argv[i] = NULL; |
| } |
| else if (strcmp ("--seed", argv[i]) == 0 || strncmp ("--seed=", argv[i], 7) == 0) |
| { |
| gchar *equal = argv[i] + 6; |
| if (*equal == '=') |
| test_run_seedstr = equal + 1; |
| else if (i + 1 < argc) |
| { |
| argv[i++] = NULL; |
| test_run_seedstr = argv[i]; |
| } |
| argv[i] = NULL; |
| } |
| else if (strcmp ("-?", argv[i]) == 0 || |
| strcmp ("-h", argv[i]) == 0 || |
| strcmp ("--help", argv[i]) == 0) |
| { |
| printf ("Usage:\n" |
| " %s [OPTION...]\n\n" |
| "Help Options:\n" |
| " -h, --help Show help options\n\n" |
| "Test Options:\n" |
| " --g-fatal-warnings Make all warnings fatal\n" |
| " -l List test cases available in a test executable\n" |
| " -m {perf|slow|thorough|quick} Execute tests according to mode\n" |
| " -m {undefined|no-undefined} Execute tests according to mode\n" |
| " -p TESTPATH Only start test cases matching TESTPATH\n" |
| " -s TESTPATH Skip all tests matching TESTPATH\n" |
| " --seed=SEEDSTRING Start tests with random seed SEEDSTRING\n" |
| " --debug-log debug test logging output\n" |
| " -q, --quiet Run tests quietly\n" |
| " --verbose Run tests verbosely\n", |
| argv[0]); |
| exit (0); |
| } |
| } |
| /* collapse argv */ |
| e = 1; |
| for (i = 1; i < argc; i++) |
| if (argv[i]) |
| { |
| argv[e++] = argv[i]; |
| if (i >= e) |
| argv[i] = NULL; |
| } |
| *argc_p = e; |
| } |
| |
| /** |
| * g_test_init: |
| * @argc: Address of the @argc parameter of the main() function. |
| * Changed if any arguments were handled. |
| * @argv: Address of the @argv parameter of main(). |
| * Any parameters understood by g_test_init() stripped before return. |
| * @...: %NULL-terminated list of special options. Currently the only |
| * defined option is `"no_g_set_prgname"`, which |
| * will cause g_test_init() to not call g_set_prgname(). |
| * |
| * Initialize the GLib testing framework, e.g. by seeding the |
| * test random number generator, the name for g_get_prgname() |
| * and parsing test related command line args. |
| * |
| * So far, the following arguments are understood: |
| * |
| * - `-l`: List test cases available in a test executable. |
| * - `--seed=SEED`: Provide a random seed to reproduce test |
| * runs using random numbers. |
| * - `--verbose`: Run tests verbosely. |
| * - `-q`, `--quiet`: Run tests quietly. |
| * - `-p PATH`: Execute all tests matching the given path. |
| * - `-s PATH`: Skip all tests matching the given path. |
| * This can also be used to force a test to run that would otherwise |
| * be skipped (ie, a test whose name contains "/subprocess"). |
| * - `-m {perf|slow|thorough|quick|undefined|no-undefined}`: Execute tests according to these test modes: |
| * |
| * `perf`: Performance tests, may take long and report results (off by default). |
| * |
| * `slow`, `thorough`: Slow and thorough tests, may take quite long and maximize coverage |
| * (off by default). |
| * |
| * `quick`: Quick tests, should run really quickly and give good coverage (the default). |
| * |
| * `undefined`: Tests for undefined behaviour, may provoke programming errors |
| * under g_test_trap_subprocess() or g_test_expect_message() to check |
| * that appropriate assertions or warnings are given (the default). |
| * |
| * `no-undefined`: Avoid tests for undefined behaviour |
| * |
| * - `--debug-log`: Debug test logging output. |
| * |
| * Since: 2.16 |
| */ |
| void |
| (g_test_init) (int *argc, |
| char ***argv, |
| ...) |
| { |
| static char seedstr[4 + 4 * 8 + 1]; |
| va_list args; |
| gpointer option; |
| /* make warnings and criticals fatal for all test programs */ |
| GLogLevelFlags fatal_mask = (GLogLevelFlags) g_log_set_always_fatal ((GLogLevelFlags) G_LOG_FATAL_MASK); |
| |
| fatal_mask = (GLogLevelFlags) (fatal_mask | G_LOG_LEVEL_WARNING | G_LOG_LEVEL_CRITICAL); |
| g_log_set_always_fatal (fatal_mask); |
| /* check caller args */ |
| g_return_if_fail (argc != NULL); |
| g_return_if_fail (argv != NULL); |
| g_return_if_fail (g_test_config_vars->test_initialized == FALSE); |
| mutable_test_config_vars.test_initialized = TRUE; |
| |
| va_start (args, argv); |
| while ((option = va_arg (args, char *))) |
| { |
| if (g_strcmp0 (option, "no_g_set_prgname") == 0) |
| no_g_set_prgname = TRUE; |
| } |
| va_end (args); |
| |
| /* setup random seed string */ |
| g_snprintf (seedstr, sizeof (seedstr), "R02S%08x%08x%08x%08x", g_random_int(), g_random_int(), g_random_int(), g_random_int()); |
| test_run_seedstr = seedstr; |
| |
| /* parse args, sets up mode, changes seed, etc. */ |
| parse_args (argc, argv); |
| |
| if (!g_get_prgname() && !no_g_set_prgname) |
| g_set_prgname ((*argv)[0]); |
| |
| /* sanity check */ |
| if (test_tap_log) |
| { |
| if (test_paths || test_startup_skip_count) |
| { |
| /* Not invoking every test (even if SKIPped) breaks the "1..XX" plan */ |
| g_printerr ("%s: -p and --GTestSkipCount options are incompatible with --tap\n", |
| (*argv)[0]); |
| exit (1); |
| } |
| } |
| |
| /* verify GRand reliability, needed for reliable seeds */ |
| if (1) |
| { |
| GRand *rg = g_rand_new_with_seed (0xc8c49fb6); |
| guint32 t1 = g_rand_int (rg), t2 = g_rand_int (rg), t3 = g_rand_int (rg), t4 = g_rand_int (rg); |
| /* g_print ("GRand-current: 0x%x 0x%x 0x%x 0x%x\n", t1, t2, t3, t4); */ |
| if (t1 != 0xfab39f9b || t2 != 0xb948fb0e || t3 != 0x3d31be26 || t4 != 0x43a19d66) |
| g_warning ("random numbers are not GRand-2.2 compatible, seeds may be broken (check $G_RANDOM_VERSION)"); |
| g_rand_free (rg); |
| } |
| |
| /* check rand seed */ |
| test_run_seed (test_run_seedstr); |
| |
| /* report program start */ |
| g_log_set_default_handler (gtest_default_log_handler, NULL); |
| g_test_log (G_TEST_LOG_START_BINARY, g_get_prgname(), test_run_seedstr, 0, NULL); |
| |
| test_argv0_dirname = g_path_get_dirname (test_argv0); |
| |
| /* Make sure we get the real dirname that the test was run from */ |
| if (g_str_has_suffix (test_argv0_dirname, "/.libs")) |
| { |
| gchar *tmp; |
| tmp = g_path_get_dirname (test_argv0_dirname); |
| g_free (test_argv0_dirname); |
| test_argv0_dirname = tmp; |
| } |
| |
| test_disted_files_dir = g_getenv ("G_TEST_SRCDIR"); |
| if (!test_disted_files_dir) |
| test_disted_files_dir = test_argv0_dirname; |
| |
| test_built_files_dir = g_getenv ("G_TEST_BUILDDIR"); |
| if (!test_built_files_dir) |
| test_built_files_dir = test_argv0_dirname; |
| } |
| |
| static void |
| test_run_seed (const gchar *rseed) |
| { |
| guint seed_failed = 0; |
| if (test_run_rand) |
| g_rand_free (test_run_rand); |
| test_run_rand = NULL; |
| while (strchr (" \t\v\r\n\f", *rseed)) |
| rseed++; |
| if (strncmp (rseed, "R02S", 4) == 0) /* seed for random generator 02 (GRand-2.2) */ |
| { |
| const char *s = rseed + 4; |
| if (strlen (s) >= 32) /* require 4 * 8 chars */ |
| { |
| guint32 seedarray[4]; |
| gchar *p, hexbuf[9] = { 0, }; |
| memcpy (hexbuf, s + 0, 8); |
| seedarray[0] = g_ascii_strtoull (hexbuf, &p, 16); |
| seed_failed += p != NULL && *p != 0; |
| memcpy (hexbuf, s + 8, 8); |
| seedarray[1] = g_ascii_strtoull (hexbuf, &p, 16); |
| seed_failed += p != NULL && *p != 0; |
| memcpy (hexbuf, s + 16, 8); |
| seedarray[2] = g_ascii_strtoull (hexbuf, &p, 16); |
| seed_failed += p != NULL && *p != 0; |
| memcpy (hexbuf, s + 24, 8); |
| seedarray[3] = g_ascii_strtoull (hexbuf, &p, 16); |
| seed_failed += p != NULL && *p != 0; |
| if (!seed_failed) |
| { |
| test_run_rand = g_rand_new_with_seed_array (seedarray, 4); |
| return; |
| } |
| } |
| } |
| g_error ("Unknown or invalid random seed: %s", rseed); |
| } |
| |
| /** |
| * g_test_rand_int: |
| * |
| * Get a reproducible random integer number. |
| * |
| * The random numbers generated by the g_test_rand_*() family of functions |
| * change with every new test program start, unless the --seed option is |
| * given when starting test programs. |
| * |
| * For individual test cases however, the random number generator is |
| * reseeded, to avoid dependencies between tests and to make --seed |
| * effective for all test cases. |
| * |
| * Returns: a random number from the seeded random number generator. |
| * |
| * Since: 2.16 |
| */ |
| gint32 |
| g_test_rand_int (void) |
| { |
| return g_rand_int (test_run_rand); |
| } |
| |
| /** |
| * g_test_rand_int_range: |
| * @begin: the minimum value returned by this function |
| * @end: the smallest value not to be returned by this function |
| * |
| * Get a reproducible random integer number out of a specified range, |
| * see g_test_rand_int() for details on test case random numbers. |
| * |
| * Returns: a number with @begin <= number < @end. |
| * |
| * Since: 2.16 |
| */ |
| gint32 |
| g_test_rand_int_range (gint32 begin, |
| gint32 end) |
| { |
| return g_rand_int_range (test_run_rand, begin, end); |
| } |
| |
| /** |
| * g_test_rand_double: |
| * |
| * Get a reproducible random floating point number, |
| * see g_test_rand_int() for details on test case random numbers. |
| * |
| * Returns: a random number from the seeded random number generator. |
| * |
| * Since: 2.16 |
| */ |
| double |
| g_test_rand_double (void) |
| { |
| return g_rand_double (test_run_rand); |
| } |
| |
| /** |
| * g_test_rand_double_range: |
| * @range_start: the minimum value returned by this function |
| * @range_end: the minimum value not returned by this function |
| * |
| * Get a reproducible random floating pointer number out of a specified range, |
| * see g_test_rand_int() for details on test case random numbers. |
| * |
| * Returns: a number with @range_start <= number < @range_end. |
| * |
| * Since: 2.16 |
| */ |
| double |
| g_test_rand_double_range (double range_start, |
| double range_end) |
| { |
| return g_rand_double_range (test_run_rand, range_start, range_end); |
| } |
| |
| /** |
| * g_test_timer_start: |
| * |
| * Start a timing test. Call g_test_timer_elapsed() when the task is supposed |
| * to be done. Call this function again to restart the timer. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_timer_start (void) |
| { |
| if (!test_user_timer) |
| test_user_timer = g_timer_new(); |
| test_user_stamp = 0; |
| g_timer_start (test_user_timer); |
| } |
| |
| /** |
| * g_test_timer_elapsed: |
| * |
| * Get the time since the last start of the timer with g_test_timer_start(). |
| * |
| * Returns: the time since the last start of the timer, as a double |
| * |
| * Since: 2.16 |
| */ |
| double |
| g_test_timer_elapsed (void) |
| { |
| test_user_stamp = test_user_timer ? g_timer_elapsed (test_user_timer, NULL) : 0; |
| return test_user_stamp; |
| } |
| |
| /** |
| * g_test_timer_last: |
| * |
| * Report the last result of g_test_timer_elapsed(). |
| * |
| * Returns: the last result of g_test_timer_elapsed(), as a double |
| * |
| * Since: 2.16 |
| */ |
| double |
| g_test_timer_last (void) |
| { |
| return test_user_stamp; |
| } |
| |
| /** |
| * g_test_minimized_result: |
| * @minimized_quantity: the reported value |
| * @format: the format string of the report message |
| * @...: arguments to pass to the printf() function |
| * |
| * Report the result of a performance or measurement test. |
| * The test should generally strive to minimize the reported |
| * quantities (smaller values are better than larger ones), |
| * this and @minimized_quantity can determine sorting |
| * order for test result reports. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_minimized_result (double minimized_quantity, |
| const char *format, |
| ...) |
| { |
| long double largs = minimized_quantity; |
| gchar *buffer; |
| va_list args; |
| |
| va_start (args, format); |
| buffer = g_strdup_vprintf (format, args); |
| va_end (args); |
| |
| g_test_log (G_TEST_LOG_MIN_RESULT, buffer, NULL, 1, &largs); |
| g_free (buffer); |
| } |
| |
| /** |
| * g_test_maximized_result: |
| * @maximized_quantity: the reported value |
| * @format: the format string of the report message |
| * @...: arguments to pass to the printf() function |
| * |
| * Report the result of a performance or measurement test. |
| * The test should generally strive to maximize the reported |
| * quantities (larger values are better than smaller ones), |
| * this and @maximized_quantity can determine sorting |
| * order for test result reports. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_maximized_result (double maximized_quantity, |
| const char *format, |
| ...) |
| { |
| long double largs = maximized_quantity; |
| gchar *buffer; |
| va_list args; |
| |
| va_start (args, format); |
| buffer = g_strdup_vprintf (format, args); |
| va_end (args); |
| |
| g_test_log (G_TEST_LOG_MAX_RESULT, buffer, NULL, 1, &largs); |
| g_free (buffer); |
| } |
| |
| /** |
| * g_test_message: |
| * @format: the format string |
| * @...: printf-like arguments to @format |
| * |
| * Add a message to the test report. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_message (const char *format, |
| ...) |
| { |
| gchar *buffer; |
| va_list args; |
| |
| va_start (args, format); |
| buffer = g_strdup_vprintf (format, args); |
| va_end (args); |
| |
| g_test_log (G_TEST_LOG_MESSAGE, buffer, NULL, 0, NULL); |
| g_free (buffer); |
| } |
| |
| /** |
| * g_test_bug_base: |
| * @uri_pattern: the base pattern for bug URIs |
| * |
| * Specify the base URI for bug reports. |
| * |
| * The base URI is used to construct bug report messages for |
| * g_test_message() when g_test_bug() is called. |
| * Calling this function outside of a test case sets the |
| * default base URI for all test cases. Calling it from within |
| * a test case changes the base URI for the scope of the test |
| * case only. |
| * Bug URIs are constructed by appending a bug specific URI |
| * portion to @uri_pattern, or by replacing the special string |
| * '\%s' within @uri_pattern if that is present. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_bug_base (const char *uri_pattern) |
| { |
| g_free (test_uri_base); |
| test_uri_base = g_strdup (uri_pattern); |
| } |
| |
| /** |
| * g_test_bug: |
| * @bug_uri_snippet: Bug specific bug tracker URI portion. |
| * |
| * This function adds a message to test reports that |
| * associates a bug URI with a test case. |
| * Bug URIs are constructed from a base URI set with g_test_bug_base() |
| * and @bug_uri_snippet. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_bug (const char *bug_uri_snippet) |
| { |
| char *c; |
| |
| g_return_if_fail (test_uri_base != NULL); |
| g_return_if_fail (bug_uri_snippet != NULL); |
| |
| c = strstr (test_uri_base, "%s"); |
| if (c) |
| { |
| char *b = g_strndup (test_uri_base, c - test_uri_base); |
| char *s = g_strconcat (b, bug_uri_snippet, c + 2, NULL); |
| g_free (b); |
| g_test_message ("Bug Reference: %s", s); |
| g_free (s); |
| } |
| else |
| g_test_message ("Bug Reference: %s%s", test_uri_base, bug_uri_snippet); |
| } |
| |
| /** |
| * g_test_get_root: |
| * |
| * Get the toplevel test suite for the test path API. |
| * |
| * Returns: the toplevel #GTestSuite |
| * |
| * Since: 2.16 |
| */ |
| GTestSuite* |
| g_test_get_root (void) |
| { |
| if (!test_suite_root) |
| { |
| test_suite_root = g_test_create_suite ("root"); |
| g_free (test_suite_root->name); |
| test_suite_root->name = g_strdup (""); |
| } |
| |
| return test_suite_root; |
| } |
| |
| /** |
| * g_test_run: |
| * |
| * Runs all tests under the toplevel suite which can be retrieved |
| * with g_test_get_root(). Similar to g_test_run_suite(), the test |
| * cases to be run are filtered according to test path arguments |
| * (`-p testpath` and `-s testpath`) as parsed by g_test_init(). |
| * g_test_run_suite() or g_test_run() may only be called once in a |
| * program. |
| * |
| * In general, the tests and sub-suites within each suite are run in |
| * the order in which they are defined. However, note that prior to |
| * GLib 2.36, there was a bug in the `g_test_add_*` |
| * functions which caused them to create multiple suites with the same |
| * name, meaning that if you created tests "/foo/simple", |
| * "/bar/simple", and "/foo/using-bar" in that order, they would get |
| * run in that order (since g_test_run() would run the first "/foo" |
| * suite, then the "/bar" suite, then the second "/foo" suite). As of |
| * 2.36, this bug is fixed, and adding the tests in that order would |
| * result in a running order of "/foo/simple", "/foo/using-bar", |
| * "/bar/simple". If this new ordering is sub-optimal (because it puts |
| * more-complicated tests before simpler ones, making it harder to |
| * figure out exactly what has failed), you can fix it by changing the |
| * test paths to group tests by suite in a way that will result in the |
| * desired running order. Eg, "/simple/foo", "/simple/bar", |
| * "/complex/foo-using-bar". |
| * |
| * However, you should never make the actual result of a test depend |
| * on the order that tests are run in. If you need to ensure that some |
| * particular code runs before or after a given test case, use |
| * g_test_add(), which lets you specify setup and teardown functions. |
| * |
| * If all tests are skipped, this function will return 0 if |
| * producing TAP output, or 77 (treated as "skip test" by Automake) otherwise. |
| * |
| * Returns: 0 on success, 1 on failure (assuming it returns at all), |
| * 0 or 77 if all tests were skipped with g_test_skip() |
| * |
| * Since: 2.16 |
| */ |
| int |
| g_test_run (void) |
| { |
| if (g_test_run_suite (g_test_get_root()) != 0) |
| return 1; |
| |
| /* 77 is special to Automake's default driver, but not Automake's TAP driver |
| * or Perl's prove(1) TAP driver. */ |
| if (test_tap_log) |
| return 0; |
| |
| if (test_run_count > 0 && test_run_count == test_skipped_count) |
| return 77; |
| else |
| return 0; |
| } |
| |
| /** |
| * g_test_create_case: |
| * @test_name: the name for the test case |
| * @data_size: the size of the fixture data structure |
| * @test_data: test data argument for the test functions |
| * @data_setup: (scope async): the function to set up the fixture data |
| * @data_test: (scope async): the actual test function |
| * @data_teardown: (scope async): the function to teardown the fixture data |
| * |
| * Create a new #GTestCase, named @test_name, this API is fairly |
| * low level, calling g_test_add() or g_test_add_func() is preferable. |
| * When this test is executed, a fixture structure of size @data_size |
| * will be automatically allocated and filled with zeros. Then @data_setup is |
| * called to initialize the fixture. After fixture setup, the actual test |
| * function @data_test is called. Once the test run completes, the |
| * fixture structure is torn down by calling @data_teardown and |
| * after that the memory is automatically released by the test framework. |
| * |
| * Splitting up a test run into fixture setup, test function and |
| * fixture teardown is most useful if the same fixture is used for |
| * multiple tests. In this cases, g_test_create_case() will be |
| * called with the same fixture, but varying @test_name and |
| * @data_test arguments. |
| * |
| * Returns: a newly allocated #GTestCase. |
| * |
| * Since: 2.16 |
| */ |
| GTestCase* |
| g_test_create_case (const char *test_name, |
| gsize data_size, |
| gconstpointer test_data, |
| GTestFixtureFunc data_setup, |
| GTestFixtureFunc data_test, |
| GTestFixtureFunc data_teardown) |
| { |
| GTestCase *tc; |
| |
| g_return_val_if_fail (test_name != NULL, NULL); |
| g_return_val_if_fail (strchr (test_name, '/') == NULL, NULL); |
| g_return_val_if_fail (test_name[0] != 0, NULL); |
| g_return_val_if_fail (data_test != NULL, NULL); |
| |
| tc = g_slice_new0 (GTestCase); |
| tc->name = g_strdup (test_name); |
| tc->test_data = (gpointer) test_data; |
| tc->fixture_size = data_size; |
| tc->fixture_setup = (void*) data_setup; |
| tc->fixture_test = (void*) data_test; |
| tc->fixture_teardown = (void*) data_teardown; |
| |
| return tc; |
| } |
| |
| static gint |
| find_suite (gconstpointer l, gconstpointer s) |
| { |
| const GTestSuite *suite = l; |
| const gchar *str = s; |
| |
| return strcmp (suite->name, str); |
| } |
| |
| static gint |
| find_case (gconstpointer l, gconstpointer s) |
| { |
| const GTestCase *tc = l; |
| const gchar *str = s; |
| |
| return strcmp (tc->name, str); |
| } |
| |
| /** |
| * GTestFixtureFunc: |
| * @fixture: (not nullable): the test fixture |
| * @user_data: the data provided when registering the test |
| * |
| * The type used for functions that operate on test fixtures. This is |
| * used for the fixture setup and teardown functions as well as for the |
| * testcases themselves. |
| * |
| * @user_data is a pointer to the data that was given when registering |
| * the test case. |
| * |
| * @fixture will be a pointer to the area of memory allocated by the |
| * test framework, of the size requested. If the requested size was |
| * zero then @fixture will be equal to @user_data. |
| * |
| * Since: 2.28 |
| */ |
| void |
| g_test_add_vtable (const char *testpath, |
| gsize data_size, |
| gconstpointer test_data, |
| GTestFixtureFunc data_setup, |
| GTestFixtureFunc fixture_test_func, |
| GTestFixtureFunc data_teardown) |
| { |
| gchar **segments; |
| guint ui; |
| GTestSuite *suite; |
| |
| g_return_if_fail (testpath != NULL); |
| g_return_if_fail (g_path_is_absolute (testpath)); |
| g_return_if_fail (fixture_test_func != NULL); |
| |
| suite = g_test_get_root(); |
| segments = g_strsplit (testpath, "/", -1); |
| for (ui = 0; segments[ui] != NULL; ui++) |
| { |
| const char *seg = segments[ui]; |
| gboolean islast = segments[ui + 1] == NULL; |
| if (islast && !seg[0]) |
| g_error ("invalid test case path: %s", testpath); |
| else if (!seg[0]) |
| continue; /* initial or duplicate slash */ |
| else if (!islast) |
| { |
| GSList *l; |
| GTestSuite *csuite; |
| l = g_slist_find_custom (suite->suites, seg, find_suite); |
| if (l) |
| { |
| csuite = l->data; |
| } |
| else |
| { |
| csuite = g_test_create_suite (seg); |
| g_test_suite_add_suite (suite, csuite); |
| } |
| suite = csuite; |
| } |
| else /* islast */ |
| { |
| GTestCase *tc; |
| |
| if (g_slist_find_custom (suite->cases, seg, find_case)) |
| g_error ("duplicate test case path: %s", testpath); |
| |
| tc = g_test_create_case (seg, data_size, test_data, data_setup, fixture_test_func, data_teardown); |
| g_test_suite_add (suite, tc); |
| } |
| } |
| g_strfreev (segments); |
| } |
| |
| /** |
| * g_test_fail: |
| * |
| * Indicates that a test failed. This function can be called |
| * multiple times from the same test. You can use this function |
| * if your test failed in a recoverable way. |
| * |
| * Do not use this function if the failure of a test could cause |
| * other tests to malfunction. |
| * |
| * Calling this function will not stop the test from running, you |
| * need to return from the test function yourself. So you can |
| * produce additional diagnostic messages or even continue running |
| * the test. |
| * |
| * If not called from inside a test, this function does nothing. |
| * |
| * Since: 2.30 |
| **/ |
| void |
| g_test_fail (void) |
| { |
| test_run_success = G_TEST_RUN_FAILURE; |
| } |
| |
| /** |
| * g_test_incomplete: |
| * @msg: (nullable): explanation |
| * |
| * Indicates that a test failed because of some incomplete |
| * functionality. This function can be called multiple times |
| * from the same test. |
| * |
| * Calling this function will not stop the test from running, you |
| * need to return from the test function yourself. So you can |
| * produce additional diagnostic messages or even continue running |
| * the test. |
| * |
| * If not called from inside a test, this function does nothing. |
| * |
| * Since: 2.38 |
| */ |
| void |
| g_test_incomplete (const gchar *msg) |
| { |
| test_run_success = G_TEST_RUN_INCOMPLETE; |
| g_free (test_run_msg); |
| test_run_msg = g_strdup (msg); |
| } |
| |
| /** |
| * g_test_skip: |
| * @msg: (nullable): explanation |
| * |
| * Indicates that a test was skipped. |
| * |
| * Calling this function will not stop the test from running, you |
| * need to return from the test function yourself. So you can |
| * produce additional diagnostic messages or even continue running |
| * the test. |
| * |
| * If not called from inside a test, this function does nothing. |
| * |
| * Since: 2.38 |
| */ |
| void |
| g_test_skip (const gchar *msg) |
| { |
| test_run_success = G_TEST_RUN_SKIPPED; |
| g_free (test_run_msg); |
| test_run_msg = g_strdup (msg); |
| } |
| |
| /** |
| * g_test_failed: |
| * |
| * Returns whether a test has already failed. This will |
| * be the case when g_test_fail(), g_test_incomplete() |
| * or g_test_skip() have been called, but also if an |
| * assertion has failed. |
| * |
| * This can be useful to return early from a test if |
| * continuing after a failed assertion might be harmful. |
| * |
| * The return value of this function is only meaningful |
| * if it is called from inside a test function. |
| * |
| * Returns: %TRUE if the test has failed |
| * |
| * Since: 2.38 |
| */ |
| gboolean |
| g_test_failed (void) |
| { |
| return test_run_success != G_TEST_RUN_SUCCESS; |
| } |
| |
| /** |
| * g_test_set_nonfatal_assertions: |
| * |
| * Changes the behaviour of g_assert_cmpstr(), g_assert_cmpint(), |
| * g_assert_cmpuint(), g_assert_cmphex(), g_assert_cmpfloat(), |
| * g_assert_true(), g_assert_false(), g_assert_null(), g_assert_no_error(), |
| * g_assert_error(), g_test_assert_expected_messages() and the various |
| * g_test_trap_assert_*() macros to not abort to program, but instead |
| * call g_test_fail() and continue. (This also changes the behavior of |
| * g_test_fail() so that it will not cause the test program to abort |
| * after completing the failed test.) |
| * |
| * Note that the g_assert_not_reached() and g_assert() are not |
| * affected by this. |
| * |
| * This function can only be called after g_test_init(). |
| * |
| * Since: 2.38 |
| */ |
| void |
| g_test_set_nonfatal_assertions (void) |
| { |
| if (!g_test_config_vars->test_initialized) |
| g_error ("g_test_set_nonfatal_assertions called without g_test_init"); |
| test_nonfatal_assertions = TRUE; |
| test_mode_fatal = FALSE; |
| } |
| |
| /** |
| * GTestFunc: |
| * |
| * The type used for test case functions. |
| * |
| * Since: 2.28 |
| */ |
| |
| /** |
| * g_test_add_func: |
| * @testpath: /-separated test case path name for the test. |
| * @test_func: (scope async): The test function to invoke for this test. |
| * |
| * Create a new test case, similar to g_test_create_case(). However |
| * the test is assumed to use no fixture, and test suites are automatically |
| * created on the fly and added to the root fixture, based on the |
| * slash-separated portions of @testpath. |
| * |
| * If @testpath includes the component "subprocess" anywhere in it, |
| * the test will be skipped by default, and only run if explicitly |
| * required via the `-p` command-line option or g_test_trap_subprocess(). |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_add_func (const char *testpath, |
| GTestFunc test_func) |
| { |
| g_return_if_fail (testpath != NULL); |
| g_return_if_fail (testpath[0] == '/'); |
| g_return_if_fail (test_func != NULL); |
| g_test_add_vtable (testpath, 0, NULL, NULL, (GTestFixtureFunc) test_func, NULL); |
| } |
| |
| /** |
| * GTestDataFunc: |
| * @user_data: the data provided when registering the test |
| * |
| * The type used for test case functions that take an extra pointer |
| * argument. |
| * |
| * Since: 2.28 |
| */ |
| |
| /** |
| * g_test_add_data_func: |
| * @testpath: /-separated test case path name for the test. |
| * @test_data: Test data argument for the test function. |
| * @test_func: (scope async): The test function to invoke for this test. |
| * |
| * Create a new test case, similar to g_test_create_case(). However |
| * the test is assumed to use no fixture, and test suites are automatically |
| * created on the fly and added to the root fixture, based on the |
| * slash-separated portions of @testpath. The @test_data argument |
| * will be passed as first argument to @test_func. |
| * |
| * If @testpath includes the component "subprocess" anywhere in it, |
| * the test will be skipped by default, and only run if explicitly |
| * required via the `-p` command-line option or g_test_trap_subprocess(). |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_add_data_func (const char *testpath, |
| gconstpointer test_data, |
| GTestDataFunc test_func) |
| { |
| g_return_if_fail (testpath != NULL); |
| g_return_if_fail (testpath[0] == '/'); |
| g_return_if_fail (test_func != NULL); |
| |
| g_test_add_vtable (testpath, 0, test_data, NULL, (GTestFixtureFunc) test_func, NULL); |
| } |
| |
| /** |
| * g_test_add_data_func_full: |
| * @testpath: /-separated test case path name for the test. |
| * @test_data: Test data argument for the test function. |
| * @test_func: The test function to invoke for this test. |
| * @data_free_func: #GDestroyNotify for @test_data. |
| * |
| * Create a new test case, as with g_test_add_data_func(), but freeing |
| * @test_data after the test run is complete. |
| * |
| * Since: 2.34 |
| */ |
| void |
| g_test_add_data_func_full (const char *testpath, |
| gpointer test_data, |
| GTestDataFunc test_func, |
| GDestroyNotify data_free_func) |
| { |
| g_return_if_fail (testpath != NULL); |
| g_return_if_fail (testpath[0] == '/'); |
| g_return_if_fail (test_func != NULL); |
| |
| g_test_add_vtable (testpath, 0, test_data, NULL, |
| (GTestFixtureFunc) test_func, |
| (GTestFixtureFunc) data_free_func); |
| } |
| |
| static gboolean |
| g_test_suite_case_exists (GTestSuite *suite, |
| const char *test_path) |
| { |
| GSList *iter; |
| char *slash; |
| GTestCase *tc; |
| |
| test_path++; |
| slash = strchr (test_path, '/'); |
| |
| if (slash) |
| { |
| for (iter = suite->suites; iter; iter = iter->next) |
| { |
| GTestSuite *child_suite = iter->data; |
| |
| if (!strncmp (child_suite->name, test_path, slash - test_path)) |
| if (g_test_suite_case_exists (child_suite, slash)) |
| return TRUE; |
| } |
| } |
| else |
| { |
| for (iter = suite->cases; iter; iter = iter->next) |
| { |
| tc = iter->data; |
| if (!strcmp (tc->name, test_path)) |
| return TRUE; |
| } |
| } |
| |
| return FALSE; |
| } |
| |
| /** |
| * g_test_create_suite: |
| * @suite_name: a name for the suite |
| * |
| * Create a new test suite with the name @suite_name. |
| * |
| * Returns: A newly allocated #GTestSuite instance. |
| * |
| * Since: 2.16 |
| */ |
| GTestSuite* |
| g_test_create_suite (const char *suite_name) |
| { |
| GTestSuite *ts; |
| g_return_val_if_fail (suite_name != NULL, NULL); |
| g_return_val_if_fail (strchr (suite_name, '/') == NULL, NULL); |
| g_return_val_if_fail (suite_name[0] != 0, NULL); |
| ts = g_slice_new0 (GTestSuite); |
| ts->name = g_strdup (suite_name); |
| return ts; |
| } |
| |
| /** |
| * g_test_suite_add: |
| * @suite: a #GTestSuite |
| * @test_case: a #GTestCase |
| * |
| * Adds @test_case to @suite. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_suite_add (GTestSuite *suite, |
| GTestCase *test_case) |
| { |
| g_return_if_fail (suite != NULL); |
| g_return_if_fail (test_case != NULL); |
| |
| suite->cases = g_slist_append (suite->cases, test_case); |
| } |
| |
| /** |
| * g_test_suite_add_suite: |
| * @suite: a #GTestSuite |
| * @nestedsuite: another #GTestSuite |
| * |
| * Adds @nestedsuite to @suite. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_suite_add_suite (GTestSuite *suite, |
| GTestSuite *nestedsuite) |
| { |
| g_return_if_fail (suite != NULL); |
| g_return_if_fail (nestedsuite != NULL); |
| |
| suite->suites = g_slist_append (suite->suites, nestedsuite); |
| } |
| |
| /** |
| * g_test_queue_free: |
| * @gfree_pointer: the pointer to be stored. |
| * |
| * Enqueue a pointer to be released with g_free() during the next |
| * teardown phase. This is equivalent to calling g_test_queue_destroy() |
| * with a destroy callback of g_free(). |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_queue_free (gpointer gfree_pointer) |
| { |
| if (gfree_pointer) |
| g_test_queue_destroy (g_free, gfree_pointer); |
| } |
| |
| /** |
| * g_test_queue_destroy: |
| * @destroy_func: Destroy callback for teardown phase. |
| * @destroy_data: Destroy callback data. |
| * |
| * This function enqueus a callback @destroy_func to be executed |
| * during the next test case teardown phase. This is most useful |
| * to auto destruct allocated test resources at the end of a test run. |
| * Resources are released in reverse queue order, that means enqueueing |
| * callback A before callback B will cause B() to be called before |
| * A() during teardown. |
| * |
| * Since: 2.16 |
| */ |
| void |
| g_test_queue_destroy (GDestroyNotify destroy_func, |
| gpointer destroy_data) |
| { |
| DestroyEntry *dentry; |
| |
| g_return_if_fail (destroy_func != NULL); |
| |
| dentry = g_slice_new0 (DestroyEntry); |
| dentry->destroy_func = destroy_func; |
| dentry->destroy_data = destroy_data; |
| dentry->next = test_destroy_queue; |
| test_destroy_queue = dentry; |
| } |
| |
| static gboolean |
| test_case_run (GTestCase *tc) |
| { |
| gchar *old_base = g_strdup (test_uri_base); |
| GSList **old_free_list, *filename_free_list = NULL; |
| gboolean success = G_TEST_RUN_SUCCESS; |
| |
| old_free_list = test_filename_free_list; |
| test_filename_free_list = &filename_free_list; |
| |
| if (++test_run_count <= test_startup_skip_count) |
| g_test_log (G_TEST_LOG_SKIP_CASE, test_run_name, NULL, 0, NULL); |
| else if (test_run_list) |
| { |
| g_print ("%s\n", test_run_name); |
| g_test_log (G_TEST_LOG_LIST_CASE, test_run_name, NULL, 0, NULL); |
| } |
| else |
| { |
| GTimer *test_run_timer = g_timer_new(); |
| long double largs[3]; |
| void *fixture; |
| g_test_log (G_TEST_LOG_START_CASE, test_run_name, NULL, 0, NULL); |
| test_run_forks = 0; |
| test_run_success = G_TEST_RUN_SUCCESS; |
| g_clear_pointer (&test_run_msg, g_free); |
| g_test_log_set_fatal_handler (NULL, NULL); |
| if (test_paths_skipped && g_slist_find_custom (test_paths_skipped, test_run_name, (GCompareFunc)g_strcmp0)) |
| g_test_skip ("by request (-s option)"); |
| else |
| { |
| g_timer_start (test_run_timer); |
| fixture = tc->fixture_size ? g_malloc0 (tc->fixture_size) : tc->test_data; |
| test_run_seed (test_run_seedstr); |
| if (tc->fixture_setup) |
| tc->fixture_setup (fixture, tc->test_data); |
| tc->fixture_test (fixture, tc->test_data); |
| test_trap_clear(); |
| while (test_destroy_queue) |
| { |
| DestroyEntry *dentry = test_destroy_queue; |
| test_destroy_queue = dentry->next; |
| dentry->destroy_func (dentry->destroy_data); |
| g_slice_free (DestroyEntry, dentry); |
| } |
| if (tc->fixture_teardown) |
| tc->fixture_teardown (fixture, tc->test_data); |
| if (tc->fixture_size) |
| g_free (fixture); |
| g_timer_stop (test_run_timer); |
| } |
| success = test_run_success; |
| test_run_success = G_TEST_RUN_FAILURE; |
| largs[0] = success; /* OK */ |
| largs[1] = test_run_forks; |
| largs[2] = g_timer_elapsed (test_run_timer, NULL); |
| g_test_log (G_TEST_LOG_STOP_CASE, test_run_name, test_run_msg, G_N_ELEMENTS (largs), largs); |
| g_clear_pointer (&test_run_msg, g_free); |
| g_timer_destroy (test_run_timer); |
| } |
| |
| g_slist_free_full (filename_free_list, g_free); |
| test_filename_free_list = old_free_list; |
| g_free (test_uri_base); |
| test_uri_base = old_base; |
| |
| return (success == G_TEST_RUN_SUCCESS || |
| success == G_TEST_RUN_SKIPPED); |
| } |
| |
| static gboolean |
| path_has_prefix (const char *path, |
| const char *prefix) |
| { |
| int prefix_len = strlen (prefix); |
| |
| return (strncmp (path, prefix, prefix_len) == 0 && |
| (path[prefix_len] == '\0' || |
| path[prefix_len] == '/')); |
| } |
| |
| static gboolean |
| test_should_run (const char *test_path, |
| const char *cmp_path) |
| { |
| if (strstr (test_run_name, "/subprocess")) |
| { |
| if (g_strcmp0 (test_path, cmp_path) == 0) |
| return TRUE; |
| |
| if (g_test_verbose ()) |
| g_print ("GTest: skipping: %s\n", test_run_name); |
| return FALSE; |
| } |
| |
| return !cmp_path || path_has_prefix (test_path, cmp_path); |
| } |
| |
| /* Recurse through @suite, running tests matching @path (or all tests |
| * if @path is %NULL). |
| */ |
| static int |
| g_test_run_suite_internal (GTestSuite *suite, |
| const char *path) |
| { |
| guint n_bad = 0; |
| gchar *old_name = test_run_name; |
| GSList *iter; |
| |
| g_return_val_if_fail (suite != NULL, -1); |
| |
| g_test_log (G_TEST_LOG_START_SUITE, suite->name, NULL, 0, NULL); |
| |
| for (iter = suite->cases; iter; iter = iter->next) |
| { |
| GTestCase *tc = iter->data; |
| |
| test_run_name = g_build_path ("/", old_name, tc->name, NULL); |
| if (test_should_run (test_run_name, path)) |
| { |
| if (!test_case_run (tc)) |
| n_bad++; |
| } |
| g_free (test_run_name); |
| } |
| |
| for (iter = suite->suites; iter; iter = iter->next) |
| { |
| GTestSuite *ts = iter->data; |
| |
| test_run_name = g_build_path ("/", old_name, ts->name, NULL); |
| if (!path || path_has_prefix (path, test_run_name)) |
| n_bad += g_test_run_suite_internal (ts, path); |
| g_free (test_run_name); |
| } |
| |
| test_run_name = old_name; |
| |
| g_test_log (G_TEST_LOG_STOP_SUITE, suite->name, NULL, 0, NULL); |
| |
| return n_bad; |
| } |
| |
| static int |
| g_test_suite_count (GTestSuite *suite) |
| { |
| int n = 0; |
| GSList *iter; |
| |
| g_return_val_if_fail (suite != NULL, -1); |
| |
| for (iter = suite->cases; iter; iter = iter->next) |
| { |
| GTestCase *tc = iter->data; |
| |
| if (strcmp (tc->name, "subprocess") != 0) |
| n++; |
| } |
| |
| for (iter = suite->suites; iter; iter = iter->next) |
| { |
| GTestSuite *ts = iter->data; |
| |
| if (strcmp (ts->name, "subprocess") != 0) |
| n += g_test_suite_count (ts); |
| } |
| |
| return n; |
| } |
| |
| /** |
| * g_test_run_suite: |
| * @suite: a #GTestSuite |
| * |
| * Execute the tests within @suite and all nested #GTestSuites. |
| * The test suites to be executed are filtered according to |
| * test path arguments (`-p testpath` and `-s testpath`) as parsed by |
| * g_test_init(). See the g_test_run() documentation for more |
| * information on the order that tests are run in. |
| * |
| * g_test_run_suite() or g_test_run() may only be called once |
| * in a program. |
| * |
| * Returns: 0 on success |
| * |
| * Since: 2.16 |
| */ |
| int |
| g_test_run_suite (GTestSuite *suite) |
| { |
| int n_bad = 0; |
| |
| g_return_val_if_fail (g_test_run_once == TRUE, -1); |
| |
| g_test_run_once = FALSE; |
| test_count = g_test_suite_count (suite); |
| |
| test_run_name = g_strdup_printf ("/%s", suite->name); |
| |
| if (test_paths) |
| { |
| GSList *iter; |
| |
| for (iter = test_paths; iter; iter = iter->next) |
| n_bad += g_test_run_suite_internal (suite, iter->data); |
| } |
| else |
| n_bad = g_test_run_suite_internal (suite, NULL); |
| |
| g_free (test_run_name); |
| test_run_name = NULL; |
| |
| return n_bad; |
| } |
| |
| static void |
| gtest_default_log_handler (const gchar *log_domain, |
| GLogLevelFlags log_level, |
| const gchar *message, |
| gpointer unused_data) |
| { |
| const gchar *strv[16]; |
| gboolean fatal = FALSE; |
| gchar *msg; |
| guint i = 0; |
| |
| if (log_domain) |
| { |
| strv[i++] = log_domain; |
| strv[i++] = "-"; |
| } |
| if (log_level & G_LOG_FLAG_FATAL) |
| { |
| strv[i++] = "FATAL-"; |
| fatal = TRUE; |
| } |
| if (log_level & G_LOG_FLAG_RECURSION) |
| strv[i++] = "RECURSIVE-"; |
| if (log_level & G_LOG_LEVEL_ERROR) |
| strv[i++] = "ERROR"; |
| if (log_level & G_LOG_LEVEL_CRITICAL) |
| strv[i++] = "CRITICAL"; |
| if (log_level & G_LOG_LEVEL_WARNING) |
| strv[i++] = "WARNING"; |
| if (log_level & G_LOG_LEVEL_MESSAGE) |
| strv[i++] = "MESSAGE"; |
| if (log_level & G_LOG_LEVEL_INFO) |
| strv[i++] = "INFO"; |
| if (log_level & G_LOG_LEVEL_DEBUG) |
| strv[i++] = "DEBUG"; |
| strv[i++] = ": "; |
| strv[i++] = message; |
| strv[i++] = NULL; |
| |
| msg = g_strjoinv ("", (gchar**) strv); |
| g_test_log (fatal ? G_TEST_LOG_ERROR : G_TEST_LOG_MESSAGE, msg, NULL, 0, NULL); |
| g_log_default_handler (log_domain, log_level, message, unused_data); |
| |
| g_free (msg); |
| } |
| |
| void |
| g_assertion_message (const char *domain, |
| const char *file, |
| int line, |
| const char *func, |
| const char *message) |
| { |
| char lstr[32]; |
| char *s; |
| |
| if (!message) |
| message = "code should not be reached"; |
| g_snprintf (lstr, 32, "%d", line); |
| s = g_strconcat (domain ? domain : "", domain && domain[0] ? ":" : "", |
| "ERROR:", file, ":", lstr, ":", |
| func, func[0] ? ":" : "", |
| " ", message, NULL); |
| g_printerr ("**\n%s\n", s); |
| |
| /* Don't print a fatal error indication if assertions are non-fatal, or |
| * if we are a child process that might be sharing the parent's stdout. */ |
| if (test_nonfatal_assertions || test_in_subprocess || test_in_forked_child) |
| g_test_log (G_TEST_LOG_MESSAGE, s, NULL, 0, NULL); |
| else |
| g_test_log (G_TEST_LOG_ERROR, s, NULL, 0, NULL); |
| |
| if (test_nonfatal_assertions) |
| { |
| g_free (s); |
| g_test_fail (); |
| return; |
| } |
| |
| /* store assertion message in global variable, so that it can be found in a |
| * core dump */ |
| if (__glib_assert_msg != NULL) |
| /* free the old one */ |
| free (__glib_assert_msg); |
| __glib_assert_msg = (char*) malloc (strlen (s) + 1); |
| strcpy (__glib_assert_msg, s); |
| |
| g_free (s); |
| |
| if (test_in_subprocess) |
| { |
| /* If this is a test case subprocess then it probably hit this |
| * assertion on purpose, so just exit() rather than abort()ing, |
| * to avoid triggering any system crash-reporting daemon. |
| */ |
| _exit (1); |
| } |
| else |
| g_abort (); |
| } |
| |
| /** |
| * g_assertion_message_expr: (skip) |
| * @domain: (nullable): |
| * @file: |
| * @line: |
| * @func: |
| * @expr: (nullable): |
| */ |
| void |
| g_assertion_message_expr (const char *domain, |
| const char *file, |
| int line, |
| const char *func, |
| const char *expr) |
| { |
| char *s; |
| if (!expr) |
| s = g_strdup ("code should not be reached"); |
| else |
| s = g_strconcat ("assertion failed: (", expr, ")", NULL); |
| g_assertion_message (domain, file, line, func, s); |
| g_free (s); |
| |
| /* Normally g_assertion_message() won't return, but we need this for |
| * when test_nonfatal_assertions is set, since |
| * g_assertion_message_expr() is used for always-fatal assertions. |
| */ |
| if (test_in_subprocess) |
| _exit (1); |
| else |
| g_abort (); |
| } |
| |
| void |
| g_assertion_message_cmpnum (const char *domain, |
| const char *file, |
| int line, |
| const char *func, |
| const char *expr, |
| long double arg1, |
| const char *cmp, |
| long double arg2, |
| char numtype) |
| { |
| char *s = NULL; |
| |
| switch (numtype) |
| { |
| case 'i': s = g_strdup_printf ("assertion failed (%s): (%" G_GINT64_MODIFIER "i %s %" G_GINT64_MODIFIER "i)", expr, (gint64) arg1, cmp, (gint64) arg2); break; |
| case 'x': s = g_strdup_printf ("assertion failed (%s): (0x%08" G_GINT64_MODIFIER "x %s 0x%08" G_GINT64_MODIFIER "x)", expr, (guint64) arg1, cmp, (guint64) arg2); break; |
| case 'f': s = g_strdup_printf ("assertion failed (%s): (%.9g %s %.9g)", expr, (double) arg1, cmp, (double) arg2); break; |
| /* ideally use: floats=%.7g double=%.17g */ |
| } |
| g_assertion_message (domain, file, line, func, s); |
| g_free (s); |
| } |
| |
| void |
| g_assertion_message_cmpstr (const char *domain, |
| const char *file, |
| int line, |
| const char *func, |
| const char *expr, |
| const char *arg1, |
| const char *cmp, |
| const char *arg2) |
| { |
| char *a1, *a2, *s, *t1 = NULL, *t2 = NULL; |
| a1 = arg1 ? g_strconcat ("\"", t1 = g_strescape (arg1, NULL), "\"", NULL) : g_strdup ("NULL"); |
| a2 = arg2 ? g_strconcat ("\"", t2 = g_strescape (arg2, NULL), "\"", NULL) : g_strdup ("NULL"); |
| g_free (t1); |
| g_free (t2); |
| s = g_strdup_printf ("assertion failed (%s): (%s %s %s)", expr, a1, cmp, a2); |
| g_free (a1); |
| g_free (a2); |
| g_assertion_message (domain, file, line, func, s); |
| g_free (s); |
| } |
| |
| void |
| g_assertion_message_error (const char *domain, |
| const char *file, |
| int line, |
| const char *func, |
| const char *expr, |
| const GError *error, |
| GQuark error_domain, |
| int error_code) |
| { |
| GString *gstring; |
| |
| /* This is used by both g_assert_error() and g_assert_no_error(), so there |
| * are three cases: expected an error but got the wrong error, expected |
| * an error but got no error, and expected no error but got an error. |
| */ |
| |
| gstring = g_string_new ("assertion failed "); |
| if (error_domain) |
| g_string_append_printf (gstring, "(%s == (%s, %d)): ", expr, |
| g_quark_to_string (error_domain), error_code); |
| else |
| g_string_append_printf (gstring, "(%s == NULL): ", expr); |
| |
| if (error) |
| g_string_append_printf (gstring, "%s (%s, %d)", error->message, |
| g_quark_to_string (error->domain), error->code); |
| else |
| g_string_append_printf (gstring, "%s is NULL", expr); |
| |
| g_assertion_message (domain, file, line, func, gstring->str); |
| g_string_free (gstring, TRUE); |
| } |
| |
| /** |
| * g_strcmp0: |
| * @str1: (nullable): a C string or %NULL |
| * @str2: (nullable): another C string or %NULL |
| * |
| * Compares @str1 and @str2 like strcmp(). Handles %NULL |
| * gracefully by sorting it before non-%NULL strings. |
| * Comparing two %NULL pointers returns 0. |
| * |
| * Returns: an integer less than, equal to, or greater than zero, if @str1 is <, == or > than @str2. |
| * |
| * Since: 2.16 |
| */ |
| int |
| g_strcmp0 (const char *str1, |
| const char *str2) |
| { |
| if (!str1) |
| return -(str1 != str2); |
| if (!str2) |
| return str1 != str2; |
| return strcmp (str1, str2); |
| } |
| |
| static void |
| test_trap_clear (void) |
| { |
| test_trap_last_status = 0; |
| test_trap_last_pid = 0; |
| g_clear_pointer (&test_trap_last_subprocess, g_free); |
| g_clear_pointer (&test_trap_last_stdout, g_free); |
| g_clear_pointer (&test_trap_last_stderr, g_free); |
| } |
| |
| #ifdef G_OS_UNIX |
| |
| static int |
| sane_dup2 (int fd1, |
| int fd2) |
| { |
| int ret; |
| do |
| ret = dup2 (fd1, fd2); |
| while (ret < 0 && errno == EINTR); |
| return ret; |
| } |
| |
| #endif |
| |
| typedef struct { |
| GPid pid; |
| GMainLoop *loop; |
| int child_status; /* unmodified platform-specific status */ |
| |
| GIOChannel *stdout_io; |
| gboolean echo_stdout; |
| GString *stdout_str; |
| |
| GIOChannel *stderr_io; |
| gboolean echo_stderr; |
| GString *stderr_str; |
| } WaitForChildData; |
| |
| static void |
| check_complete (WaitForChildData *data) |
| { |
| if (data->child_status != -1 && data->stdout_io == NULL && data->stderr_io == NULL) |
| g_main_loop_quit (data->loop); |
| } |
| |
| static void |
| child_exited (GPid pid, |
| gint status, |
| gpointer user_data) |
| { |
| WaitForChildData *data = user_data; |
| |
| g_assert (status != -1); |
| data->child_status = status; |
| |
| check_complete (data); |
| } |
| |
| static gboolean |
| child_timeout (gpointer user_data) |
| { |
| WaitForChildData *data = user_data; |
| |
| #ifdef G_OS_WIN32 |
| TerminateProcess (data->pid, G_TEST_STATUS_TIMED_OUT); |
| #else |
| kill (data->pid, SIGALRM); |
| #endif |
| |
| return FALSE; |
| } |
| |
| static gboolean |
| child_read (GIOChannel *io, GIOCondition cond, gpointer user_data) |
| { |
| WaitForChildData *data = user_data; |
| GIOStatus status; |
| gsize nread, nwrote, total; |
| gchar buf[4096]; |
| FILE *echo_file = NULL; |
| |
| status = g_io_channel_read_chars (io, buf, sizeof (buf), &nread, NULL); |
| if (status == G_IO_STATUS_ERROR || status == G_IO_STATUS_EOF) |
| { |
| // FIXME data->error = (status == G_IO_STATUS_ERROR); |
| if (io == data->stdout_io) |
| g_clear_pointer (&data->stdout_io, g_io_channel_unref); |
| else |
| g_clear_pointer (&data->stderr_io, g_io_channel_unref); |
| |
| check_complete (data); |
| return FALSE; |
| } |
| else if (status == G_IO_STATUS_AGAIN) |
| return TRUE; |
| |
| if (io == data->stdout_io) |
| { |
| g_string_append_len (data->stdout_str, buf, nread); |
| if (data->echo_stdout) |
| echo_file = stdout; |
| } |
| else |
| { |
| g_string_append_len (data->stderr_str, buf, nread); |
| if (data->echo_stderr) |
| echo_file = stderr; |
| } |
| |
| if (echo_file) |
| { |
| for (total = 0; total < nread; total += nwrote) |
| { |
| int errsv; |
| |
| nwrote = fwrite (buf + total, 1, nread - total, echo_file); |
| errsv = errno; |
| if (nwrote == 0) |
| g_error ("write failed: %s", g_strerror (errsv)); |
| } |
| } |
| |
| return TRUE; |
| } |
| |
| static void |
| wait_for_child (GPid pid, |
| int stdout_fd, gboolean echo_stdout, |
| int stderr_fd, gboolean echo_stderr, |
| guint64 timeout) |
| { |
| WaitForChildData data; |
| GMainContext *context; |
| GSource *source; |
| |
| data.pid = pid; |
| data.child_status = -1; |
| |
| context = g_main_context_new (); |
| data.loop = g_main_loop_new (context, FALSE); |
| |
| source = g_child_watch_source_new (pid); |
| g_source_set_callback (source, (GSourceFunc) child_exited, &data, NULL); |
| g_source_attach (source, context); |
| g_source_unref (source); |
| |
| data.echo_stdout = echo_stdout; |
| data.stdout_str = g_string_new (NULL); |
| data.stdout_io = g_io_channel_unix_new (stdout_fd); |
| g_io_channel_set_close_on_unref (data.stdout_io, TRUE); |
| g_io_channel_set_encoding (data.stdout_io, NULL, NULL); |
| g_io_channel_set_buffered (data.stdout_io, FALSE); |
| source = g_io_create_watch (data.stdout_io, G_IO_IN | G_IO_ERR | G_IO_HUP); |
| g_source_set_callback (source, (GSourceFunc) child_read, &data, NULL); |
| g_source_attach (source, context); |
| g_source_unref (source); |
| |
| data.echo_stderr = echo_stderr; |
| data.stderr_str = g_string_new (NULL); |
| data.stderr_io = g_io_channel_unix_new (stderr_fd); |
| g_io_channel_set_close_on_unref (data.stderr_io, TRUE); |
| g_io_channel_set_encoding (data.stderr_io, NULL, NULL); |
| g_io_channel_set_buffered (data.stderr_io, FALSE); |
| source = g_io_create_watch (data.stderr_io, G_IO_IN | G_IO_ERR | G_IO_HUP); |
| g_source_set_callback (source, (GSourceFunc) child_read, &data, NULL); |
| g_source_attach (source, context); |
| g_source_unref (source); |
| |
| if (timeout) |
| { |
| source = g_timeout_source_new (0); |
| g_source_set_ready_time (source, g_get_monotonic_time () + timeout); |
| g_source_set_callback (source, (GSourceFunc) child_timeout, &data, NULL); |
| g_source_attach (source, context); |
| g_source_unref (source); |
| } |
| |
| g_main_loop_run (data.loop); |
| g_main_loop_unref (data.loop); |
| g_main_context_unref (context); |
| |
| test_trap_last_pid = pid; |
| test_trap_last_status = data.child_status; |
| test_trap_last_stdout = g_string_free (data.stdout_str, FALSE); |
| test_trap_last_stderr = g_string_free (data.stderr_str, FALSE); |
| |
| g_clear_pointer (&data.stdout_io, g_io_channel_unref); |
| g_clear_pointer (&data.stderr_io, g_io_channel_unref); |
| } |
| |
| /** |
| * g_test_trap_fork: |
| * @usec_timeout: Timeout for the forked test in micro seconds. |
| * @test_trap_flags: Flags to modify forking behaviour. |
| * |
| * Fork the current test program to execute a test case that might |
| * not return or that might abort. |
| * |
| * If @usec_timeout is non-0, the forked test case is aborted and |
| * considered failing if its run time exceeds it. |
| * |
| * The forking behavior can be configured with the #GTestTrapFlags flags. |
| * |
| * In the following example, the test code forks, the forked child |
| * process produces some sample output and exits successfully. |
| * The forking parent process then asserts successful child program |
| * termination and validates child program outputs. |
| * |
| * |[<!-- language="C" --> |
| * static void |
| * test_fork_patterns (void) |
| * { |
| * if (g_test_trap_fork (0, G_TEST_TRAP_SILENCE_STDOUT | G_TEST_TRAP_SILENCE_STDERR)) |
| * { |
| * g_print ("some stdout text: somagic17\n"); |
| * g_printerr ("some stderr text: semagic43\n"); |
| * exit (0); // successful test run |
| * } |
| * g_test_trap_assert_passed (); |
| * g_test_trap_assert_stdout ("*somagic17*"); |
| * g_test_trap_assert_stderr ("*semagic43*"); |
| * } |
| * ]| |
| * |
| * Returns: %TRUE for the forked child and %FALSE for the executing parent process. |
| * |
| * Since: 2.16 |
| * |
| * Deprecated: This function is implemented only on Unix platforms, |
| * and is not always reliable due to problems inherent in |
| * fork-without-exec. Use g_test_trap_subprocess() instead. |
| */ |
| gboolean |
| g_test_trap_fork (guint64 usec_timeout, |
| GTestTrapFlags test_trap_flags) |
| { |
| #ifdef G_OS_UNIX |
| int stdout_pipe[2] = { -1, -1 }; |
| int stderr_pipe[2] = { -1, -1 }; |
| int errsv; |
| |
| test_trap_clear(); |
| if (pipe (stdout_pipe) < 0 || pipe (stderr_pipe) < 0) |
| { |
| errsv = errno; |
| g_error ("failed to create pipes to fork test program: %s", g_strerror (errsv)); |
| } |
| test_trap_last_pid = fork (); |
| errsv = errno; |
| if (test_trap_last_pid < 0) |
| g_error ("failed to fork test program: %s", g_strerror (errsv)); |
| if (test_trap_last_pid == 0) /* child */ |
| { |
| int fd0 = -1; |
| test_in_forked_child = TRUE; |
| close (stdout_pipe[0]); |
| close (stderr_pipe[0]); |
| if (!(test_trap_flags & G_TEST_TRAP_INHERIT_STDIN)) |
| { |
| fd0 = g_open ("/dev/null", O_RDONLY, 0); |
| if (fd0 < 0) |
| g_error ("failed to open /dev/null for stdin redirection"); |
| } |
| if (sane_dup2 (stdout_pipe[1], 1) < 0 || sane_dup2 (stderr_pipe[1], 2) < 0 || (fd0 >= 0 && sane_dup2 (fd0, 0) < 0)) |
| { |
| errsv = errno; |
| g_error ("failed to dup2() in forked test program: %s", g_strerror (errsv)); |
| } |
| if (fd0 >= 3) |
| close (fd0); |
| if (stdout_pipe[1] >= 3) |
| close (stdout_pipe[1]); |
| if (stderr_pipe[1] >= 3) |
| close (stderr_pipe[1]); |
| return TRUE; |
| } |
| else /* parent */ |
| { |
| test_run_forks++; |
| close (stdout_pipe[1]); |
| close (stderr_pipe[1]); |
| |
| wait_for_child (test_trap_last_pid, |
| stdout_pipe[0], !(test_trap_flags & G_TEST_TRAP_SILENCE_STDOUT), |
| stderr_pipe[0], !(test_trap_flags & G_TEST_TRAP_SILENCE_STDERR), |
| usec_timeout); |
| return FALSE; |
| } |
| #else |
| g_message ("Not implemented: g_test_trap_fork"); |
| |
| return FALSE; |
| #endif |
| } |
| |
| /** |
| * g_test_trap_subprocess: |
| * @test_path: (nullable): Test to run in a subprocess |
| * @usec_timeout: Timeout for the subprocess test in micro seconds. |
| * @test_flags: Flags to modify subprocess behaviour. |
| * |
| * Respawns the test program to run only @test_path in a subprocess. |
| * This can be used for a test case that might not return, or that |
| * might abort. |
| * |
| * If @test_path is %NULL then the same test is re-run in a subprocess. |
| * You can use g_test_subprocess() to determine whether the test is in |
| * a subprocess or not. |
| * |
| * @test_path can also be the name of the parent test, followed by |
| * "`/subprocess/`" and then a name for the specific subtest (or just |
| * ending with "`/subprocess`" if the test only has one child test); |
| * tests with names of this form will automatically be skipped in the |
| * parent process. |
| * |
| * If @usec_timeout is non-0, the test subprocess is aborted and |
| * considered failing if its run time exceeds it. |
| * |
| * The subprocess behavior can be configured with the |
| * #GTestSubprocessFlags flags. |
| * |
| * You can use methods such as g_test_trap_assert_passed(), |
| * g_test_trap_assert_failed(), and g_test_trap_assert_stderr() to |
| * check the results of the subprocess. (But note that |
| * g_test_trap_assert_stdout() and g_test_trap_assert_stderr() |
| * cannot be used if @test_flags specifies that the child should |
| * inherit the parent stdout/stderr.) |
| * |
| * If your `main ()` needs to behave differently in |
| * the subprocess, you can call g_test_subprocess() (after calling |
| * g_test_init()) to see whether you are in a subprocess. |
| * |
| * The following example tests that calling |
| * `my_object_new(1000000)` will abort with an error |
| * message. |
| * |
| * |[<!-- language="C" --> |
| * static void |
| * test_create_large_object (void) |
| * { |
| * if (g_test_subprocess ()) |
| * { |
| * my_object_new (1000000); |
| * return; |
| * } |
| * |
| * // Reruns this same test in a subprocess |
| * g_test_trap_subprocess (NULL, 0, 0); |
| * g_test_trap_assert_failed (); |
| * g_test_trap_assert_stderr ("*ERROR*too large*"); |
| * } |
| * |
| * int |
| * main (int argc, char **argv) |
| * { |
| * g_test_init (&argc, &argv, NULL); |
| * |
| * g_test_add_func ("/myobject/create_large_object", |
| * test_create_large_object); |
| * return g_test_run (); |
| * } |
| * ]| |
| * |
| * Since: 2.38 |
| */ |
| void |
| g_test_trap_subprocess (const char *test_path, |
| guint64 usec_timeout, |
| GTestSubprocessFlags test_flags) |
| { |
| GError *error = NULL; |
| GPtrArray *argv; |
| GSpawnFlags flags; |
| int stdout_fd, stderr_fd; |
| GPid pid; |
| |
| /* Sanity check that they used GTestSubprocessFlags, not GTestTrapFlags */ |
| g_assert ((test_flags & (G_TEST_TRAP_INHERIT_STDIN | G_TEST_TRAP_SILENCE_STDOUT | G_TEST_TRAP_SILENCE_STDERR)) == 0); |
| |
| if (test_path) |
| { |
| if (!g_test_suite_case_exists (g_test_get_root (), test_path)) |
| g_error ("g_test_trap_subprocess: test does not exist: %s", test_path); |
| } |
| else |
| { |
| test_path = test_run_name; |
| } |
| |
| if (g_test_verbose ()) |
| g_print ("GTest: subprocess: %s\n", test_path); |
| |
| test_trap_clear (); |
| test_trap_last_subprocess = g_strdup (test_path); |
| |
| argv = g_ptr_array_new (); |
| g_ptr_array_add (argv, test_argv0); |
| g_ptr_array_add (argv, "-q"); |
| g_ptr_array_add (argv, "-p"); |
| g_ptr_array_add (argv, (char *)test_path); |
| g_ptr_array_add (argv, "--GTestSubprocess"); |
| if (test_log_fd != -1) |
| { |
| char log_fd_buf[128]; |
| |
| g_ptr_array_add (argv, "--GTestLogFD"); |
| g_snprintf (log_fd_buf, sizeof (log_fd_buf), "%d", test_log_fd); |
| g_ptr_array_add (argv, log_fd_buf); |
| } |
| g_ptr_array_add (argv, NULL); |
| |
| flags = G_SPAWN_DO_NOT_REAP_CHILD; |
| if (test_flags & G_TEST_TRAP_INHERIT_STDIN) |
| flags |= G_SPAWN_CHILD_INHERITS_STDIN; |
| |
| if (!g_spawn_async_with_pipes (test_initial_cwd, |
| (char **)argv->pdata, |
| NULL, flags, |
| NULL, NULL, |
| &pid, NULL, &stdout_fd, &stderr_fd, |
| &error)) |
| { |
| g_error ("g_test_trap_subprocess() failed: %s", |
| error->message); |
| } |
| g_ptr_array_free (argv, TRUE); |
| |
| wait_for_child (pid, |
| stdout_fd, !!(test_flags & G_TEST_SUBPROCESS_INHERIT_STDOUT), |
| stderr_fd, !!(test_flags & G_TEST_SUBPROCESS_INHERIT_STDERR), |
| usec_timeout); |
| } |
| |
| /** |
| * g_test_subprocess: |
| * |
| * Returns %TRUE (after g_test_init() has been called) if the test |
| * program is running under g_test_trap_subprocess(). |
| * |
| * Returns: %TRUE if the test program is running under |
| * g_test_trap_subprocess(). |
| * |
| * Since: 2.38 |
| */ |
| gboolean |
| g_test_subprocess (void) |
| { |
| return test_in_subprocess; |
| } |
| |
| /** |
| * g_test_trap_has_passed: |
| * |
| * Check the result of the last g_test_trap_subprocess() call. |
| * |
| * Returns: %TRUE if the last test subprocess terminated successfully. |
| * |
| * Since: 2.16 |
| */ |
| gboolean |
| g_test_trap_has_passed (void) |
| { |
| #ifdef G_OS_UNIX |
| return (WIFEXITED (test_trap_last_status) && |
| WEXITSTATUS (test_trap_last_status) == 0); |
| #else |
| return test_trap_last_status == 0; |
| #endif |
| } |
| |
| /** |
| * g_test_trap_reached_timeout: |
| * |
| * Check the result of the last g_test_trap_subprocess() call. |
| * |
| * Returns: %TRUE if the last test subprocess got killed due to a timeout. |
| * |
| * Since: 2.16 |
| */ |
| gboolean |
| g_test_trap_reached_timeout (void) |
| { |
| #ifdef G_OS_UNIX |
| return (WIFSIGNALED (test_trap_last_status) && |
| WTERMSIG (test_trap_last_status) == SIGALRM); |
| #else |
| return test_trap_last_status == G_TEST_STATUS_TIMED_OUT; |
| #endif |
| } |
| |
| static gboolean |
| log_child_output (const gchar *process_id) |
| { |
| gchar *escaped; |
| |
| #ifdef G_OS_UNIX |
| if (WIFEXITED (test_trap_last_status)) /* normal exit */ |
| { |
| if (WEXITSTATUS (test_trap_last_status) == 0) |
| g_test_message ("child process (%s) exit status: 0 (success)", |
| process_id); |
| else |
| g_test_message ("child process (%s) exit status: %d (error)", |
| process_id, WEXITSTATUS (test_trap_last_status)); |
| } |
| else if (WIFSIGNALED (test_trap_last_status) && |
| WTERMSIG (test_trap_last_status) == SIGALRM) |
| { |
| g_test_message ("child process (%s) timed out", process_id); |
| } |
| else if (WIFSIGNALED (test_trap_last_status)) |
| { |
| const gchar *maybe_dumped_core = ""; |
| |
| #ifdef WCOREDUMP |
| if (WCOREDUMP (test_trap_last_status)) |
| maybe_dumped_core = ", core dumped"; |
| #endif |
| |
| g_test_message ("child process (%s) killed by signal %d (%s)%s", |
| process_id, WTERMSIG (test_trap_last_status), |
| g_strsignal (WTERMSIG (test_trap_last_status)), |
| maybe_dumped_core); |
| } |
| else |
| { |
| g_test_message ("child process (%s) unknown wait status %d", |
| process_id, test_trap_last_status); |
| } |
| #else |
| if (test_trap_last_status == 0) |
| g_test_message ("child process (%s) exit status: 0 (success)", |
| process_id); |
| else |
| g_test_message ("child process (%s) exit status: %d (error)", |
| process_id, test_trap_last_status); |
| #endif |
| |
| escaped = g_strescape (test_trap_last_stdout, NULL); |
| g_test_message ("child process (%s) stdout: \"%s\"", process_id, escaped); |
| g_free (escaped); |
| |
| escaped = g_strescape (test_trap_last_stderr, NULL); |
| g_test_message ("child process (%s) stderr: \"%s\"", process_id, escaped); |
| g_free (escaped); |
| |
| /* so we can use short-circuiting: |
| * logged_child_output = logged_child_output || log_child_output (...) */ |
| return TRUE; |
| } |
| |
| void |
| g_test_trap_assertions (const char *domain, |
| const char *file, |
| int line, |
| const char *func, |
| guint64 assertion_flags, /* 0-pass, 1-fail, 2-outpattern, 4-errpattern */ |
| const char *pattern) |
| { |
| gboolean must_pass = assertion_flags == 0; |
| gboolean must_fail = assertion_flags == 1; |
| gboolean match_result = 0 == (assertion_flags & 1); |
| gboolean logged_child_output = FALSE; |
| const char *stdout_pattern = (assertion_flags & 2) ? pattern : NULL; |
| const char *stderr_pattern = (assertion_flags & 4) ? pattern : NULL; |
| const char *match_error = match_result ? "failed to match" : "contains invalid match"; |
| char *process_id; |
| |
| #ifdef G_OS_UNIX |
| if (test_trap_last_subprocess != NULL) |
| { |
| process_id = g_strdup_printf ("%s [%d]", test_trap_last_subprocess, |
| test_trap_last_pid); |
| } |
| else if (test_trap_last_pid != 0) |
| process_id = g_strdup_printf ("%d", test_trap_last_pid); |
| #else |
| if (test_trap_last_subprocess != NULL) |
| process_id = g_strdup (test_trap_last_subprocess); |
| #endif |
| else |
| g_error ("g_test_trap_ assertion with no trapped test"); |
| |
| if (must_pass && !g_test_trap_has_passed()) |
| { |
| char *msg; |
| |
| logged_child_output = logged_child_output || log_child_output (process_id); |
| |
| msg = g_strdup_printf ("child process (%s) failed unexpectedly", process_id); |
| g_assertion_message (domain, file, line, func, msg); |
| g_free (msg); |
| } |
| if (must_fail && g_test_trap_has_passed()) |
| { |
| char *msg; |
| |
| logged_child_output = logged_child_output || log_child_output (process_id); |
| |
| msg = g_strdup_printf ("child process (%s) did not fail as expected", process_id); |
| g_assertion_message (domain, file, line, func, msg); |
| g_free (msg); |
| } |
| if (stdout_pattern && match_result == !g_pattern_match_simple (stdout_pattern, test_trap_last_stdout)) |
| { |
| char *msg; |
| |
| logged_child_output = logged_child_output || log_child_output (process_id); |
| |
| msg = g_strdup_printf ("stdout of child process (%s) %s: %s", process_id, match_error, stdout_pattern); |
| g_assertion_message (domain, file, line, func, msg); |
| g_free (msg); |
| } |
| if (stderr_pattern && match_result == !g_pattern_match_simple (stderr_pattern, test_trap_last_stderr)) |
| { |
| char *msg; |
| |
| logged_child_output = logged_child_output || log_child_output (process_id); |
| |
| msg = g_strdup_printf ("stderr of child process (%s) %s: %s", process_id, match_error, stderr_pattern); |
| g_assertion_message (domain, file, line, func, msg); |
| g_free (msg); |
| } |
| g_free (process_id); |
| } |
| |
| static void |
| gstring_overwrite_int (GString *gstring, |
| guint pos, |
| guint32 vuint) |
| { |
| vuint = g_htonl (vuint); |
| g_string_overwrite_len (gstring, pos, (const gchar*) &vuint, 4); |
| } |
| |
| static void |
| gstring_append_int (GString *gstring, |
| guint32 vuint) |
| { |
| vuint = g_htonl (vuint); |
| g_string_append_len (gstring, (const gchar*) &vuint, 4); |
| } |
| |
| static void |
| gstring_append_double (GString *gstring, |
| double vdouble) |
| { |
| union { double vdouble; guint64 vuint64; } u; |
| u.vdouble = vdouble; |
| u.vuint64 = GUINT64_TO_BE (u.vuint64); |
| g_string_append_len (gstring, (const gchar*) &u.vuint64, 8); |
| } |
| |
| static guint8* |
| g_test_log_dump (GTestLogMsg *msg, |
| guint *len) |
| { |
| GString *gstring = g_string_sized_new (1024); |
| guint ui; |
| gstring_append_int (gstring, 0); /* message length */ |
| gstring_append_int (gstring, msg->log_type); |
| gstring_append_int (gstring, msg->n_strings); |
| gstring_append_int (gstring, msg->n_nums); |
| gstring_append_int (gstring, 0); /* reserved */ |
| for (ui = 0; ui < msg->n_strings; ui++) |
| { |
| guint l = strlen (msg->strings[ui]); |
| gstring_append_int (gstring, l); |
| g_string_append_len (gstring, msg->strings[ui], l); |
| } |
| for (ui = 0; ui < msg->n_nums; ui++) |
| gstring_append_double (gstring, msg->nums[ui]); |
| *len = gstring->len; |
| gstring_overwrite_int (gstring, 0, *len); /* message length */ |
| return (guint8*) g_string_free (gstring, FALSE); |
| } |
| |
| static inline long double |
| net_double (const gchar **ipointer) |
| { |
| union { guint64 vuint64; double vdouble; } u; |
| guint64 aligned_int64; |
| memcpy (&aligned_int64, *ipointer, 8); |
| *ipointer += 8; |
| u.vuint64 = GUINT64_FROM_BE (aligned_int64); |
| return u.vdouble; |
| } |
| |
| static inline guint32 |
| |