blob: e695979bb3fbff419400177ac7cb8bcaa00fd635 [file] [log] [blame] [edit]
#!/usr/bin/env python3
# Copyright 2016 The PDFium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import argparse
from dataclasses import dataclass, field
import multiprocessing
import os
import re
import shutil
import subprocess
import sys
import time
import common
import pdfium_root
import pngdiffer
from skia_gold import skia_gold
import suppressor
pdfium_root.add_source_directory_to_import_path(os.path.join('build', 'util'))
from lib.results import result_sink, result_types
# Arbitrary timestamp, expressed in seconds since the epoch, used to make sure
# that tests that depend on the current time are stable. Happens to be the
# timestamp of the first commit to repo, 2014/5/9 17:48:50.
TEST_SEED_TIME = "1399672130"
# List of test types that should run text tests instead of pixel tests.
TEXT_TESTS = ['javascript']
class TestRunner:
def __init__(self, dirname):
# Currently the only used directories are corpus, javascript, and pixel,
# which all correspond directly to the type for the test being run. In the
# future if there are tests that don't have this clean correspondence, then
# an argument for the type will need to be added.
self.per_process_config = _PerProcessConfig(
test_dir=dirname, test_type=dirname)
@property
def options(self):
return self.per_process_config.options
def IsSkiaGoldEnabled(self):
return (self.options.run_skia_gold and
not self.per_process_config.test_type in TEXT_TESTS)
def IsExecutionSuppressed(self, input_path):
return self.per_process_state.test_suppressor.IsExecutionSuppressed(
input_path)
def IsResultSuppressed(self, input_filename):
return self.per_process_state.test_suppressor.IsResultSuppressed(
input_filename)
def HandleResult(self, test_case, test_result):
input_filename = os.path.basename(test_case.input_path)
test_result.status = self._SuppressStatus(input_filename,
test_result.status)
if test_result.status == result_types.UNKNOWN:
self.result_suppressed_cases.append(input_filename)
self.surprises.append(test_case.input_path)
elif test_result.status == result_types.SKIP:
self.result_suppressed_cases.append(input_filename)
elif not test_result.IsPass():
self.failures.append(test_case.input_path)
for artifact in test_result.image_artifacts:
if artifact.skia_gold_status == result_types.PASS:
if self.IsResultSuppressed(artifact.image_path):
self.skia_gold_unexpected_successes.append(artifact.GetSkiaGoldId())
else:
self.skia_gold_successes.append(artifact.GetSkiaGoldId())
elif artifact.skia_gold_status == result_types.FAIL:
self.skia_gold_failures.append(artifact.GetSkiaGoldId())
# Log test result.
print(f'{test_result.status}: {test_result.test_id}')
if not test_result.IsPass():
if test_result.reason:
print(f'Failure reason: {test_result.reason}')
if test_result.log:
print(f'Test output:\n{test_result.log}')
for artifact in test_result.image_artifacts:
if artifact.skia_gold_status == result_types.FAIL:
print(f'Failed Skia Gold: {artifact.image_path}')
if artifact.image_diff:
print(f'Failed image diff: {artifact.image_diff.reason}')
# Report test result to ResultDB.
if self.resultdb:
only_artifacts = None
if len(test_result.image_artifacts) == 1:
only_artifacts = test_result.image_artifacts[0].GetDiffArtifacts()
self.resultdb.Post(
test_id=test_result.test_id,
status=test_result.status,
duration=test_result.duration_milliseconds,
test_log=test_result.log,
test_file=None,
artifacts=only_artifacts,
failure_reason=test_result.reason)
# Milo only supports a single diff per test, so if we have multiple pages,
# report each page as its own "test."
if len(test_result.image_artifacts) > 1:
for page, artifact in enumerate(test_result.image_artifacts):
self.resultdb.Post(
test_id=f'{test_result.test_id}/{page}',
status=self._SuppressArtifactStatus(test_result,
artifact.GetDiffStatus()),
duration=None,
test_log=None,
test_file=None,
artifacts=artifact.GetDiffArtifacts(),
failure_reason=artifact.GetDiffReason())
def _SuppressStatus(self, input_filename, status):
if not self.IsResultSuppressed(input_filename):
return status
if status == result_types.PASS:
# There isn't an actual status for succeeded-but-ignored, so use the
# "abort" status to differentiate this from failed-but-ignored.
#
# Note that this appears as a preliminary failure in Gerrit.
return result_types.UNKNOWN
# There isn't an actual status for failed-but-ignored, so use the "skip"
# status to differentiate this from succeeded-but-ignored.
return result_types.SKIP
def _SuppressArtifactStatus(self, test_result, status):
if status != result_types.FAIL:
return status
if test_result.status != result_types.SKIP:
return status
return result_types.SKIP
def Run(self):
# Running a test defines a number of attributes on the fly.
# pylint: disable=attribute-defined-outside-init
relative_test_dir = self.per_process_config.test_dir
if relative_test_dir != 'corpus':
relative_test_dir = os.path.join('resources', relative_test_dir)
parser = argparse.ArgumentParser()
parser.add_argument(
'--build-dir',
default=os.path.join('out', 'Debug'),
help='relative path from the base source directory')
parser.add_argument(
'-j',
default=multiprocessing.cpu_count(),
dest='num_workers',
type=int,
help='run NUM_WORKERS jobs in parallel')
parser.add_argument(
'--disable-javascript',
action="store_true",
dest="disable_javascript",
help='Prevents JavaScript from executing in PDF files.')
parser.add_argument(
'--disable-xfa',
action="store_true",
dest="disable_xfa",
help='Prevents processing XFA forms.')
parser.add_argument(
'--render-oneshot',
action="store_true",
dest="render_oneshot",
help='Sets whether to use the oneshot renderer.')
parser.add_argument(
'--run-skia-gold',
action='store_true',
default=False,
help='When flag is on, skia gold tests will be run.')
# TODO: Remove when pdfium recipe stops passing this argument
parser.add_argument(
'--gold_properties',
default='',
dest="gold_properties",
help='Key value pairs that are written to the top level '
'of the JSON file that is ingested by Gold.')
# TODO: Remove when pdfium recipe stops passing this argument
parser.add_argument(
'--gold_ignore_hashes',
default='',
dest="gold_ignore_hashes",
help='Path to a file with MD5 hashes we wish to ignore.')
parser.add_argument(
'--regenerate_expected',
default='',
dest="regenerate_expected",
help='Regenerates expected images. Valid values are '
'"all" to regenerate all expected pngs, and '
'"platform" to regenerate only platform-specific '
'expected pngs.')
parser.add_argument(
'--reverse-byte-order',
action='store_true',
dest="reverse_byte_order",
help='Run image-based tests using --reverse-byte-order.')
parser.add_argument(
'--ignore_errors',
action="store_true",
dest="ignore_errors",
help='Prevents the return value from being non-zero '
'when image comparison fails.')
parser.add_argument(
'inputted_file_paths',
nargs='*',
help='Path to test files to run, relative to '
f'testing/{relative_test_dir}. If omitted, runs all test files under '
f'testing/{relative_test_dir}.',
metavar='relative/test/path')
skia_gold.add_skia_gold_args(parser)
self.per_process_config.options = parser.parse_args()
if (self.options.regenerate_expected and
self.options.regenerate_expected not in ['all', 'platform']):
print('FAILURE: --regenerate_expected must be "all" or "platform"')
return 1
finder = self.per_process_config.NewFinder()
pdfium_test_path = self.per_process_config.GetPdfiumTestPath(finder)
if not os.path.exists(pdfium_test_path):
print(f"FAILURE: Can't find test executable '{pdfium_test_path}'")
print('Use --build-dir to specify its location.')
return 1
self.per_process_config.InitializeFeatures(pdfium_test_path)
self.per_process_state = _PerProcessState(self.per_process_config)
shutil.rmtree(self.per_process_state.working_dir, ignore_errors=True)
os.makedirs(self.per_process_state.working_dir)
error_message = self.per_process_state.image_differ.CheckMissingTools(
self.options.regenerate_expected)
if error_message:
print('FAILURE:', error_message)
return 1
self.resultdb = result_sink.TryInitClient()
if self.resultdb:
print('Detected ResultSink environment')
# Collect test cases.
walk_from_dir = finder.TestingDir(relative_test_dir)
self.test_cases = TestCaseManager()
self.execution_suppressed_cases = []
input_file_re = re.compile('^.+[.](in|pdf)$')
if self.options.inputted_file_paths:
for file_name in self.options.inputted_file_paths:
input_path = os.path.join(walk_from_dir, file_name)
if not os.path.isfile(input_path):
print(f"Can't find test file '{file_name}'")
return 1
self.test_cases.NewTestCase(input_path)
else:
for file_dir, _, filename_list in os.walk(walk_from_dir):
for input_filename in filename_list:
if input_file_re.match(input_filename):
input_path = os.path.join(file_dir, input_filename)
if self.IsExecutionSuppressed(input_path):
self.execution_suppressed_cases.append(input_path)
continue
if not os.path.isfile(input_path):
continue
self.test_cases.NewTestCase(input_path)
# Execute test cases.
self.failures = []
self.surprises = []
self.skia_gold_successes = []
self.skia_gold_unexpected_successes = []
self.skia_gold_failures = []
self.result_suppressed_cases = []
if self.IsSkiaGoldEnabled():
assert self.options.gold_output_dir
# Clear out and create top level gold output directory before starting
skia_gold.clear_gold_output_dir(self.options.gold_output_dir)
with multiprocessing.Pool(
processes=self.options.num_workers,
initializer=_InitializePerProcessState,
initargs=[self.per_process_config]) as pool:
if self.per_process_config.test_type in TEXT_TESTS:
test_function = _RunTextTest
else:
test_function = _RunPixelTest
for result in pool.imap(test_function, self.test_cases):
self.HandleResult(self.test_cases.GetTestCase(result.test_id), result)
# Report test results.
if self.surprises:
self.surprises.sort()
print('\nUnexpected Successes:')
for surprise in self.surprises:
print(surprise)
if self.failures:
self.failures.sort()
print('\nSummary of Failures:')
for failure in self.failures:
print(failure)
if self.skia_gold_unexpected_successes:
self.skia_gold_unexpected_successes.sort()
print('\nUnexpected Skia Gold Successes:')
for surprise in self.skia_gold_unexpected_successes:
print(surprise)
if self.skia_gold_failures:
self.skia_gold_failures.sort()
print('\nSummary of Skia Gold Failures:')
for failure in self.skia_gold_failures:
print(failure)
self._PrintSummary()
if self.failures:
if not self.options.ignore_errors:
return 1
return 0
def _PrintSummary(self):
number_test_cases = len(self.test_cases)
number_failures = len(self.failures)
number_suppressed = len(self.result_suppressed_cases)
number_successes = number_test_cases - number_failures - number_suppressed
number_surprises = len(self.surprises)
print('\nTest cases executed:', number_test_cases)
print(' Successes:', number_successes)
print(' Suppressed:', number_suppressed)
print(' Surprises:', number_surprises)
print(' Failures:', number_failures)
if self.IsSkiaGoldEnabled():
number_gold_failures = len(self.skia_gold_failures)
number_gold_successes = len(self.skia_gold_successes)
number_gold_surprises = len(self.skia_gold_unexpected_successes)
number_total_gold_tests = sum(
[number_gold_failures, number_gold_successes, number_gold_surprises])
print('\nSkia Gold Test cases executed:', number_total_gold_tests)
print(' Skia Gold Successes:', number_gold_successes)
print(' Skia Gold Surprises:', number_gold_surprises)
print(' Skia Gold Failures:', number_gold_failures)
skia_tester = self.per_process_state.GetSkiaGoldTester()
if self.skia_gold_failures and skia_tester.IsTryjobRun():
cl_triage_link = skia_tester.GetCLTriageLink()
print(' Triage link for CL:', cl_triage_link)
skia_tester.WriteCLTriageLink(cl_triage_link)
print()
print('Test cases not executed:', len(self.execution_suppressed_cases))
def SetDeleteOutputOnSuccess(self, new_value):
"""Set whether to delete generated output if the test passes."""
self.per_process_config.delete_output_on_success = new_value
def SetEnforceExpectedImages(self, new_value):
"""Set whether to enforce that each test case provide an expected image."""
self.per_process_config.enforce_expected_images = new_value
def _RunTextTest(test_case):
"""Runs a text test case."""
test_case_runner = _TestCaseRunner(test_case)
with test_case_runner:
test_case_runner.test_result = test_case_runner.GenerateAndTest(
test_case_runner.TestText)
return test_case_runner.test_result
def _RunPixelTest(test_case):
"""Runs a pixel test case."""
test_case_runner = _TestCaseRunner(test_case)
with test_case_runner:
test_case_runner.test_result = test_case_runner.GenerateAndTest(
test_case_runner.TestPixel)
return test_case_runner.test_result
# `_PerProcessState` singleton. This is initialized when creating the
# `multiprocessing.Pool()`. `TestRunner.Run()` creates its own separate
# instance of `_PerProcessState` as well.
_per_process_state = None
def _InitializePerProcessState(config):
"""Initializes the `_per_process_state` singleton."""
global _per_process_state
assert not _per_process_state
_per_process_state = _PerProcessState(config)
@dataclass
class _PerProcessConfig:
"""Configuration for initializing `_PerProcessState`.
Attributes:
test_dir: The name of the test directory.
test_type: The test type.
delete_output_on_success: Whether to delete output on success.
enforce_expected_images: Whether to enforce expected images.
options: The dictionary of command line options.
features: The list of features supported by `pdfium_test`.
"""
test_dir: str
test_type: str
delete_output_on_success: bool = False
enforce_expected_images: bool = False
options: dict = None
features: list = None
def NewFinder(self):
return common.DirectoryFinder(self.options.build_dir)
def GetPdfiumTestPath(self, finder):
return finder.ExecutablePath('pdfium_test')
def InitializeFeatures(self, pdfium_test_path):
output = subprocess.check_output([pdfium_test_path, '--show-config'])
self.features = output.decode('utf-8').strip().split(',')
class _PerProcessState:
"""State defined per process."""
def __init__(self, config):
self.test_dir = config.test_dir
self.test_type = config.test_type
self.delete_output_on_success = config.delete_output_on_success
self.enforce_expected_images = config.enforce_expected_images
self.options = config.options
self.features = config.features
finder = config.NewFinder()
self.pdfium_test_path = config.GetPdfiumTestPath(finder)
self.fixup_path = finder.ScriptPath('fixup_pdf_template.py')
self.text_diff_path = finder.ScriptPath('text_diff.py')
self.font_dir = os.path.join(finder.TestingDir(), 'resources', 'fonts')
self.third_party_font_dir = finder.ThirdPartyFontsDir()
self.source_dir = finder.TestingDir()
self.working_dir = finder.WorkingDir(os.path.join('testing', self.test_dir))
self.test_suppressor = suppressor.Suppressor(
finder, self.features, self.options.disable_javascript,
self.options.disable_xfa)
self.image_differ = pngdiffer.PNGDiffer(finder, self.features,
self.options.reverse_byte_order)
self.process_name = multiprocessing.current_process().name
self.skia_tester = None
def __getstate__(self):
raise RuntimeError('Cannot pickle per-process state')
def GetSkiaGoldTester(self):
"""Gets the `SkiaGoldTester` singleton for this worker."""
if not self.skia_tester:
self.skia_tester = skia_gold.SkiaGoldTester(
source_type=self.test_type,
skia_gold_args=self.options,
process_name=self.process_name)
return self.skia_tester
class _TestCaseRunner:
"""Runner for a single test case."""
def __init__(self, test_case):
self.test_case = test_case
self.test_result = None
self.duration_start = 0
self.source_dir, self.input_filename = os.path.split(
self.test_case.input_path)
self.pdf_path = os.path.join(self.working_dir, f'{self.test_id}.pdf')
self.actual_images = None
def __enter__(self):
self.duration_start = time.perf_counter_ns()
return self
def __exit__(self, exc_type, exc_value, traceback):
if not self.test_result:
self.test_result = self.test_case.NewResult(
result_types.UNKNOWN, reason='No test result recorded')
duration = time.perf_counter_ns() - self.duration_start
self.test_result.duration_milliseconds = duration * 1e-6
@property
def options(self):
return _per_process_state.options
@property
def test_id(self):
return self.test_case.test_id
@property
def working_dir(self):
return _per_process_state.working_dir
def IsResultSuppressed(self):
return _per_process_state.test_suppressor.IsResultSuppressed(
self.input_filename)
def IsImageDiffSuppressed(self):
return _per_process_state.test_suppressor.IsImageDiffSuppressed(
self.input_filename)
def RunCommand(self, command, stdout=None):
"""Runs a test command.
Args:
command: The list of command arguments.
stdout: Optional `file`-like object to send standard output.
Returns:
The test result.
"""
if stdout:
stderr = subprocess.PIPE
else:
stdout = subprocess.PIPE
stderr = subprocess.STDOUT
completed_process = subprocess.run(
command, stdout=stdout, stderr=stderr, check=False, encoding='utf-8')
if completed_process.returncode != 0:
if stdout == subprocess.PIPE:
test_log = completed_process.stdout
else:
test_log = completed_process.stderr
return self.test_case.NewResult(
result_types.FAIL,
log=test_log,
reason='Command {} exited with code {}'.format(
completed_process.args, completed_process.returncode))
return self.test_case.NewResult(result_types.PASS)
def GenerateAndTest(self, test_function):
"""Generate test input and run pdfium_test."""
test_result = self.Generate()
if not test_result.IsPass():
return test_result
return test_function()
# TODO(crbug.com/pdfium/1508): Add support for an option to automatically
# generate Skia/SkiaPaths specific expected results.
def _RegenerateIfNeeded(self):
if not self.options.regenerate_expected:
return
if self.IsResultSuppressed() or self.IsImageDiffSuppressed():
return
_per_process_state.image_differ.Regenerate(
self.input_filename,
self.source_dir,
self.working_dir,
platform_only=self.options.regenerate_expected == 'platform')
def Generate(self):
input_event_path = os.path.join(self.source_dir, f'{self.test_id}.evt')
if os.path.exists(input_event_path):
output_event_path = f'{os.path.splitext(self.pdf_path)[0]}.evt'
shutil.copyfile(input_event_path, output_event_path)
template_path = os.path.join(self.source_dir, f'{self.test_id}.in')
if not os.path.exists(template_path):
if os.path.exists(self.test_case.input_path):
shutil.copyfile(self.test_case.input_path, self.pdf_path)
return self.test_case.NewResult(result_types.PASS)
return self.RunCommand([
sys.executable, _per_process_state.fixup_path,
f'--output-dir={self.working_dir}', template_path
])
def TestText(self):
txt_path = os.path.join(self.working_dir, f'{self.test_id}.txt')
with open(txt_path, 'w') as outfile:
cmd_to_run = [
_per_process_state.pdfium_test_path, '--send-events',
f'--time={TEST_SEED_TIME}'
]
if self.options.disable_javascript:
cmd_to_run.append('--disable-javascript')
if self.options.disable_xfa:
cmd_to_run.append('--disable-xfa')
cmd_to_run.append(self.pdf_path)
test_result = self.RunCommand(cmd_to_run, stdout=outfile)
if not test_result.IsPass():
return test_result
# If the expected file does not exist, the output is expected to be empty.
expected_txt_path = os.path.join(self.source_dir,
f'{self.test_id}_expected.txt')
if not os.path.exists(expected_txt_path):
return self._VerifyEmptyText(txt_path)
# If JavaScript is disabled, the output should be empty.
# However, if the test is suppressed and JavaScript is disabled, do not
# verify that the text is empty so the suppressed test does not surprise.
if self.options.disable_javascript and not self.IsResultSuppressed():
return self._VerifyEmptyText(txt_path)
return self.RunCommand([
sys.executable, _per_process_state.text_diff_path, expected_txt_path,
txt_path
])
def _VerifyEmptyText(self, txt_path):
with open(txt_path, "r") as txt_file:
txt_data = txt_file.read()
if txt_data:
return self.test_case.NewResult(
result_types.FAIL, log=txt_data, reason=f'{txt_path} should be empty')
return self.test_case.NewResult(result_types.PASS)
# TODO(crbug.com/pdfium/1656): Remove when ready to fully switch over to
# Skia Gold
def TestPixel(self):
# Remove any existing generated images from previous runs.
self.actual_images = _per_process_state.image_differ.GetActualFiles(
self.input_filename, self.source_dir, self.working_dir)
self._CleanupPixelTest()
# Generate images.
cmd_to_run = [
_per_process_state.pdfium_test_path, '--send-events', '--png', '--md5',
f'--time={TEST_SEED_TIME}'
]
if 'use_ahem' in self.source_dir or 'use_symbolneu' in self.source_dir:
cmd_to_run.append(f'--font-dir={_per_process_state.font_dir}')
else:
cmd_to_run.append(f'--font-dir={_per_process_state.third_party_font_dir}')
cmd_to_run.append('--croscore-font-names')
if self.options.disable_javascript:
cmd_to_run.append('--disable-javascript')
if self.options.disable_xfa:
cmd_to_run.append('--disable-xfa')
if self.options.render_oneshot:
cmd_to_run.append('--render-oneshot')
if self.options.reverse_byte_order:
cmd_to_run.append('--reverse-byte-order')
cmd_to_run.append(self.pdf_path)
raised_exception, results = common.RunCommandExtractHashedFiles(cmd_to_run)
if raised_exception:
return self.test_case.NewResult(
result_types.FAIL, reason=str(raised_exception))
test_result = self.test_case.NewResult(
result_types.PASS,
image_artifacts=[
self._NewImageArtifact(image_path=image_path, md5_hash=md5_hash)
for image_path, md5_hash in results
])
if self.actual_images:
image_diffs = _per_process_state.image_differ.ComputeDifferences(
self.input_filename, self.source_dir, self.working_dir)
if image_diffs:
test_result.status = result_types.FAIL
test_result.reason = 'Images differ'
# Merge image diffs into test result.
diff_map = {}
diff_log = []
for diff in image_diffs:
diff_map[diff.actual_path] = diff
diff_log.append((f'{os.path.basename(diff.actual_path)} vs. '
f'{os.path.basename(diff.expected_path)}\n'))
for artifact in test_result.image_artifacts:
artifact.image_diff = diff_map.get(artifact.image_path)
test_result.log = ''.join(diff_log)
elif _per_process_state.enforce_expected_images:
if not self.IsImageDiffSuppressed():
test_result.status = result_types.FAIL
test_result.reason = 'Missing expected images'
if not test_result.IsPass():
self._RegenerateIfNeeded()
return test_result
if _per_process_state.delete_output_on_success:
self._CleanupPixelTest()
return test_result
def _NewImageArtifact(self, *, image_path, md5_hash):
artifact = ImageArtifact(image_path=image_path, md5_hash=md5_hash)
if self.options.run_skia_gold:
if _per_process_state.GetSkiaGoldTester().UploadTestResultToSkiaGold(
artifact.GetSkiaGoldId(), artifact.image_path):
artifact.skia_gold_status = result_types.PASS
else:
artifact.skia_gold_status = result_types.FAIL
return artifact
def _CleanupPixelTest(self):
for image_file in self.actual_images:
if os.path.exists(image_file):
os.remove(image_file)
@dataclass
class TestCase:
"""Description of a test case to run.
Attributes:
test_id: A unique identifier for the test.
input_path: The absolute path to the test file.
"""
test_id: str
input_path: str
def NewResult(self, status, **kwargs):
"""Derives a new test result corresponding to this test case."""
return TestResult(test_id=self.test_id, status=status, **kwargs)
@dataclass
class TestResult:
"""Results from running a test case.
Attributes:
test_id: The corresponding test case ID.
status: The overall `result_types` status.
duration_milliseconds: Test time in milliseconds.
log: Optional log of the test's output.
image_artfacts: Optional list of image artifacts.
reason: Optional reason why the test failed.
"""
test_id: str
status: str
duration_milliseconds: float = None
log: str = None
image_artifacts: list = field(default_factory=list)
reason: str = None
def IsPass(self):
"""Whether the test passed."""
return self.status == result_types.PASS
@dataclass
class ImageArtifact:
"""Image artifact for a test result.
Attributes:
image_path: The absolute path to the image file.
md5_hash: The MD5 hash of the pixel buffer.
skia_gold_status: Optional Skia Gold status.
image_diff: Optional image diff.
"""
image_path: str
md5_hash: str
skia_gold_status: str = None
image_diff: pngdiffer.ImageDiff = None
def GetSkiaGoldId(self):
# The output filename without image extension becomes the test ID. For
# example, "/path/to/.../testing/corpus/example_005.pdf.0.png" becomes
# "example_005.pdf.0".
return _GetTestId(os.path.basename(self.image_path))
def GetDiffStatus(self):
return result_types.FAIL if self.image_diff else result_types.PASS
def GetDiffReason(self):
return self.image_diff.reason if self.image_diff else None
def GetDiffArtifacts(self):
if not self.image_diff:
return None
if not self.image_diff.expected_path or not self.image_diff.diff_path:
return None
return {
'actual_image':
_GetArtifactFromFilePath(self.image_path),
'expected_image':
_GetArtifactFromFilePath(self.image_diff.expected_path),
'image_diff':
_GetArtifactFromFilePath(self.image_diff.diff_path)
}
class TestCaseManager:
"""Manages a collection of test cases."""
def __init__(self):
self.test_cases = {}
def __len__(self):
return len(self.test_cases)
def __iter__(self):
return iter(self.test_cases.values())
def NewTestCase(self, input_path, **kwargs):
"""Creates and registers a new test case."""
input_basename = os.path.basename(input_path)
test_id = _GetTestId(input_basename)
if test_id in self.test_cases:
raise ValueError(
f'Test ID "{test_id}" derived from "{input_basename}" must be unique')
test_case = TestCase(test_id=test_id, input_path=input_path, **kwargs)
self.test_cases[test_id] = test_case
return test_case
def GetTestCase(self, test_id):
"""Looks up a test case previously registered by `NewTestCase()`."""
return self.test_cases[test_id]
def _GetTestId(input_basename):
"""Constructs a test ID by stripping the last extension from the basename."""
return os.path.splitext(input_basename)[0]
def _GetArtifactFromFilePath(file_path):
"""Constructs a ResultSink artifact from a file path."""
return {'filePath': file_path}