blob: cb976945202af6db397ecb32a91bc6ddbda327ee [file] [log] [blame]
#!/usr/bin/env python
# Copyright 2018, Google Inc.
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are
# met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above
# copyright notice, this list of conditions and the following disclaimer
# in the documentation and/or other materials provided with the
# distribution.
# * Neither the name of Google Inc. nor the names of its
# contributors may be used to endorse or promote products derived from
# this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
"""Unit test for the gtest_json_output module."""
import datetime
import errno
import json
import os
import re
import sys
from googletest.test import gtest_json_test_utils
from googletest.test import gtest_test_utils
GTEST_FILTER_FLAG = '--gtest_filter'
GTEST_LIST_TESTS_FLAG = '--gtest_list_tests'
GTEST_OUTPUT_FLAG = '--gtest_output'
GTEST_DEFAULT_OUTPUT_FILE = 'test_detail.json'
GTEST_PROGRAM_NAME = 'gtest_xml_output_unittest_'
# The flag indicating stacktraces are not supported
NO_STACKTRACE_SUPPORT_FLAG = '--no_stacktrace_support'
SUPPORTS_STACK_TRACES = NO_STACKTRACE_SUPPORT_FLAG not in sys.argv
if SUPPORTS_STACK_TRACES:
STACK_TRACE_TEMPLATE = '\nStack trace:\n*'
else:
STACK_TRACE_TEMPLATE = '\n'
EXPECTED_NON_EMPTY = {
'tests': 26,
'failures': 5,
'disabled': 2,
'errors': 0,
'timestamp': '*',
'time': '*',
'ad_hoc_property': '42',
'name': 'AllTests',
'testsuites': [
{
'name': 'SuccessfulTest',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'Succeeds',
'file': 'gtest_xml_output_unittest_.cc',
'line': 53,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'SuccessfulTest',
}],
},
{
'name': 'FailedTest',
'tests': 1,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'Fails',
'file': 'gtest_xml_output_unittest_.cc',
'line': 61,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'FailedTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
},
{
'name': 'DisabledTest',
'tests': 1,
'failures': 0,
'disabled': 1,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'DISABLED_test_not_run',
'file': 'gtest_xml_output_unittest_.cc',
'line': 68,
'status': 'NOTRUN',
'result': 'SUPPRESSED',
'time': '*',
'timestamp': '*',
'classname': 'DisabledTest',
}],
},
{
'name': 'SkippedTest',
'tests': 3,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'Skipped',
'file': 'gtest_xml_output_unittest_.cc',
'line': 75,
'status': 'RUN',
'result': 'SKIPPED',
'time': '*',
'timestamp': '*',
'classname': 'SkippedTest',
},
{
'name': 'SkippedWithMessage',
'file': 'gtest_xml_output_unittest_.cc',
'line': 79,
'status': 'RUN',
'result': 'SKIPPED',
'time': '*',
'timestamp': '*',
'classname': 'SkippedTest',
},
{
'name': 'SkippedAfterFailure',
'file': 'gtest_xml_output_unittest_.cc',
'line': 83,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'SkippedTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
},
],
},
{
'name': 'MixedResultTest',
'tests': 3,
'failures': 1,
'disabled': 1,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'Succeeds',
'file': 'gtest_xml_output_unittest_.cc',
'line': 88,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'MixedResultTest',
},
{
'name': 'Fails',
'file': 'gtest_xml_output_unittest_.cc',
'line': 93,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'MixedResultTest',
'failures': [
{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
},
{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 2\n 3'
+ STACK_TRACE_TEMPLATE
),
'type': '',
},
],
},
{
'name': 'DISABLED_test',
'file': 'gtest_xml_output_unittest_.cc',
'line': 98,
'status': 'NOTRUN',
'result': 'SUPPRESSED',
'time': '*',
'timestamp': '*',
'classname': 'MixedResultTest',
},
],
},
{
'name': 'XmlQuotingTest',
'tests': 1,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'OutputsCData',
'file': 'gtest_xml_output_unittest_.cc',
'line': 102,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'XmlQuotingTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Failed\nXML output: <?xml encoding="utf-8">'
'<top><![CDATA[cdata text]]></top>'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
},
{
'name': 'InvalidCharactersTest',
'tests': 1,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'InvalidCharactersInMessage',
'file': 'gtest_xml_output_unittest_.cc',
'line': 109,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'InvalidCharactersTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Failed\nInvalid characters in brackets'
' [\x01\x02]'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
},
{
'name': 'PropertyRecordingTest',
'tests': 4,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'SetUpTestSuite': 'yes',
'TearDownTestSuite': 'aye',
'testsuite': [
{
'name': 'OneProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 121,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_1': '1',
},
{
'name': 'IntValuedProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 125,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_int': '1',
},
{
'name': 'ThreeProperties',
'file': 'gtest_xml_output_unittest_.cc',
'line': 129,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_1': '1',
'key_2': '2',
'key_3': '3',
},
{
'name': 'TwoValuesForOneKeyUsesLastValue',
'file': 'gtest_xml_output_unittest_.cc',
'line': 135,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_1': '2',
},
],
},
{
'name': 'NoFixtureTest',
'tests': 3,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'RecordProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 140,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'NoFixtureTest',
'key': '1',
},
{
'name': 'ExternalUtilityThatCallsRecordIntValuedProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 153,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'NoFixtureTest',
'key_for_utility_int': '1',
},
{
'name': (
'ExternalUtilityThatCallsRecordStringValuedProperty'
),
'file': 'gtest_xml_output_unittest_.cc',
'line': 157,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'NoFixtureTest',
'key_for_utility_string': '1',
},
],
},
{
'name': 'TypedTest/0',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'int',
'file': 'gtest_xml_output_unittest_.cc',
'line': 173,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'TypedTest/0',
}],
},
{
'name': 'TypedTest/1',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'long',
'file': 'gtest_xml_output_unittest_.cc',
'line': 173,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'TypedTest/1',
}],
},
{
'name': 'Single/TypeParameterizedTestSuite/0',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'int',
'file': 'gtest_xml_output_unittest_.cc',
'line': 180,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/TypeParameterizedTestSuite/0',
}],
},
{
'name': 'Single/TypeParameterizedTestSuite/1',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'long',
'file': 'gtest_xml_output_unittest_.cc',
'line': 180,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/TypeParameterizedTestSuite/1',
}],
},
{
'name': 'Single/ValueParamTest',
'tests': 4,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'HasValueParamAttribute/0',
'value_param': '33',
'file': 'gtest_xml_output_unittest_.cc',
'line': 164,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
{
'name': 'HasValueParamAttribute/1',
'value_param': '42',
'file': 'gtest_xml_output_unittest_.cc',
'line': 164,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
{
'name': 'AnotherTestThatHasValueParamAttribute/0',
'value_param': '33',
'file': 'gtest_xml_output_unittest_.cc',
'line': 165,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
{
'name': 'AnotherTestThatHasValueParamAttribute/1',
'value_param': '42',
'file': 'gtest_xml_output_unittest_.cc',
'line': 165,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
],
},
],
}
EXPECTED_FILTERED = {
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'name': 'AllTests',
'ad_hoc_property': '42',
'testsuites': [{
'name': 'SuccessfulTest',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'Succeeds',
'file': 'gtest_xml_output_unittest_.cc',
'line': 53,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'SuccessfulTest',
}],
}],
}
EXPECTED_NO_TEST = {
'tests': 0,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'name': 'AllTests',
'testsuites': [{
'name': 'NonTestSuiteFailure',
'tests': 1,
'failures': 1,
'disabled': 0,
'skipped': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': '',
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': '',
'failures': [{
'failure': (
'gtest_no_test_unittest.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
}],
}
GTEST_PROGRAM_PATH = gtest_test_utils.GetTestExecutablePath(GTEST_PROGRAM_NAME)
SUPPORTS_TYPED_TESTS = (
'TypedTest'
in gtest_test_utils.Subprocess(
[GTEST_PROGRAM_PATH, GTEST_LIST_TESTS_FLAG], capture_stderr=False
).output
)
class GTestJsonOutputUnitTest(gtest_test_utils.TestCase):
"""Unit test for Google Test's JSON output functionality."""
# This test currently breaks on platforms that do not support typed and
# type-parameterized tests, so we don't run it under them.
if SUPPORTS_TYPED_TESTS:
def testNonEmptyJsonOutput(self):
"""Verifies JSON output for a Google Test binary with non-empty output.
Runs a test program that generates a non-empty JSON output, and
tests that the JSON output is expected.
"""
self._TestJsonOutput(GTEST_PROGRAM_NAME, EXPECTED_NON_EMPTY, 1)
def testNoTestJsonOutput(self):
"""Verifies JSON output for a Google Test binary without actual tests.
Runs a test program that generates an JSON output for a binary with no
tests, and tests that the JSON output is expected.
"""
self._TestJsonOutput('gtest_no_test_unittest', EXPECTED_NO_TEST, 0)
def testTimestampValue(self):
"""Checks whether the timestamp attribute in the JSON output is valid.
Runs a test program that generates an empty JSON output, and checks if
the timestamp attribute in the testsuites tag is valid.
"""
actual = self._GetJsonOutput('gtest_no_test_unittest', [], 0)
date_time_str = actual['timestamp']
# datetime.strptime() is only available in Python 2.5+ so we have to
# parse the expected datetime manually.
match = re.match(r'(\d+)-(\d\d)-(\d\d)T(\d\d):(\d\d):(\d\d)', date_time_str)
self.assertTrue(
re.match,
'JSON datettime string %s has incorrect format' % date_time_str,
)
date_time_from_json = datetime.datetime(
year=int(match.group(1)),
month=int(match.group(2)),
day=int(match.group(3)),
hour=int(match.group(4)),
minute=int(match.group(5)),
second=int(match.group(6)),
)
time_delta = abs(datetime.datetime.now() - date_time_from_json)
# timestamp value should be near the current local time
self.assertTrue(
time_delta < datetime.timedelta(seconds=600),
'time_delta is %s' % time_delta,
)
def testDefaultOutputFile(self):
"""Verifies the default output file name.
Confirms that Google Test produces an JSON output file with the expected
default name if no name is explicitly specified.
"""
output_file = os.path.join(
gtest_test_utils.GetTempDir(), GTEST_DEFAULT_OUTPUT_FILE
)
gtest_prog_path = gtest_test_utils.GetTestExecutablePath(
'gtest_no_test_unittest'
)
try:
os.remove(output_file)
except OSError:
e = sys.exc_info()[1]
if e.errno != errno.ENOENT:
raise
p = gtest_test_utils.Subprocess(
[gtest_prog_path, '%s=json' % GTEST_OUTPUT_FLAG],
working_dir=gtest_test_utils.GetTempDir(),
)
self.assertTrue(p.exited)
self.assertEqual(0, p.exit_code)
self.assertTrue(os.path.isfile(output_file))
def testSuppressedJsonOutput(self):
"""Verifies that no JSON output is generated.
Tests that no JSON file is generated if the default JSON listener is
shut down before RUN_ALL_TESTS is invoked.
"""
json_path = os.path.join(
gtest_test_utils.GetTempDir(), GTEST_PROGRAM_NAME + 'out.json'
)
if os.path.isfile(json_path):
os.remove(json_path)
command = [
GTEST_PROGRAM_PATH,
'%s=json:%s' % (GTEST_OUTPUT_FLAG, json_path),
'--shut_down_xml',
]
p = gtest_test_utils.Subprocess(command)
if p.terminated_by_signal:
# p.signal is available only if p.terminated_by_signal is True.
self.assertFalse(
p.terminated_by_signal,
'%s was killed by signal %d' % (GTEST_PROGRAM_NAME, p.signal),
)
else:
self.assertTrue(p.exited)
self.assertEqual(
1,
p.exit_code,
"'%s' exited with code %s, which doesn't match "
'the expected exit code %s.' % (command, p.exit_code, 1),
)
self.assertTrue(not os.path.isfile(json_path))
def testFilteredTestJsonOutput(self):
"""Verifies JSON output when a filter is applied.
Runs a test program that executes only some tests and verifies that
non-selected tests do not show up in the JSON output.
"""
self._TestJsonOutput(
GTEST_PROGRAM_NAME,
EXPECTED_FILTERED,
0,
extra_args=['%s=SuccessfulTest.*' % GTEST_FILTER_FLAG],
)
def _GetJsonOutput(self, gtest_prog_name, extra_args, expected_exit_code):
"""Returns the JSON output generated by running the program gtest_prog_name.
Furthermore, the program's exit code must be expected_exit_code.
Args:
gtest_prog_name: Google Test binary name.
extra_args: extra arguments to binary invocation.
expected_exit_code: program's exit code.
"""
json_path = os.path.join(
gtest_test_utils.GetTempDir(), gtest_prog_name + 'out.json'
)
gtest_prog_path = gtest_test_utils.GetTestExecutablePath(gtest_prog_name)
command = [
gtest_prog_path,
'%s=json:%s' % (GTEST_OUTPUT_FLAG, json_path),
] + extra_args
p = gtest_test_utils.Subprocess(command)
if p.terminated_by_signal:
self.assertTrue(
False, '%s was killed by signal %d' % (gtest_prog_name, p.signal)
)
else:
self.assertTrue(p.exited)
self.assertEqual(
expected_exit_code,
p.exit_code,
"'%s' exited with code %s, which doesn't match "
'the expected exit code %s.'
% (command, p.exit_code, expected_exit_code),
)
with open(json_path) as f:
actual = json.load(f)
return actual
def _TestJsonOutput(
self, gtest_prog_name, expected, expected_exit_code, extra_args=None
):
"""Checks the JSON output generated by the Google Test binary.
Asserts that the JSON document generated by running the program
gtest_prog_name matches expected_json, a string containing another
JSON document. Furthermore, the program's exit code must be
expected_exit_code.
Args:
gtest_prog_name: Google Test binary name.
expected: expected output.
expected_exit_code: program's exit code.
extra_args: extra arguments to binary invocation.
"""
actual = self._GetJsonOutput(
gtest_prog_name, extra_args or [], expected_exit_code
)
self.assertEqual(expected, gtest_json_test_utils.normalize(actual))
if __name__ == '__main__':
if NO_STACKTRACE_SUPPORT_FLAG in sys.argv:
# unittest.main() can't handle unknown flags
sys.argv.remove(NO_STACKTRACE_SUPPORT_FLAG)
os.environ['GTEST_STACK_TRACE_DEPTH'] = '1'
gtest_test_utils.Main()