123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542 |
- #!/usr/bin/python
- # Script to compare testsuite failures against a list of known-to-fail
- # tests.
- #
- # NOTE: This script is used in installations that are running Python 2.4.
- # Please stick to syntax features available in 2.4 and earlier
- # versions.
- # Contributed by Diego Novillo <dnovillo@google.com>
- #
- # Copyright (C) 2011-2013 Free Software Foundation, Inc.
- #
- # This file is part of GCC.
- #
- # GCC is free software; you can redistribute it and/or modify
- # it under the terms of the GNU General Public License as published by
- # the Free Software Foundation; either version 3, or (at your option)
- # any later version.
- #
- # GCC is distributed in the hope that it will be useful,
- # but WITHOUT ANY WARRANTY; without even the implied warranty of
- # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- # GNU General Public License for more details.
- #
- # You should have received a copy of the GNU General Public License
- # along with GCC; see the file COPYING. If not, write to
- # the Free Software Foundation, 51 Franklin Street, Fifth Floor,
- # Boston, MA 02110-1301, USA.
- """This script provides a coarser XFAILing mechanism that requires no
- detailed DejaGNU markings. This is useful in a variety of scenarios:
- - Development branches with many known failures waiting to be fixed.
- - Release branches with known failures that are not considered
- important for the particular release criteria used in that branch.
- The script must be executed from the toplevel build directory. When
- executed it will:
- 1- Determine the target built: TARGET
- 2- Determine the source directory: SRCDIR
- 3- Look for a failure manifest file in
- <SRCDIR>/<MANIFEST_SUBDIR>/<MANIFEST_NAME>.xfail
- 4- Collect all the <tool>.sum files from the build tree.
- 5- Produce a report stating:
- a- Failures expected in the manifest but not present in the build.
- b- Failures in the build not expected in the manifest.
- 6- If all the build failures are expected in the manifest, it exits
- with exit code 0. Otherwise, it exits with error code 1.
- Manifest files contain expected DejaGNU results that are otherwise
- treated as failures.
- They may also contain additional text:
- # This is a comment. - self explanatory
- @include file - the file is a path relative to the includer
- @remove result text - result text is removed from the expected set
- """
- import datetime
- import optparse
- import os
- import re
- import sys
- # Handled test results.
- _VALID_TEST_RESULTS = [ 'FAIL', 'UNRESOLVED', 'XPASS', 'ERROR' ]
- _VALID_TEST_RESULTS_REX = re.compile("%s" % "|".join(_VALID_TEST_RESULTS))
- # Subdirectory of srcdir in which to find the manifest file.
- _MANIFEST_SUBDIR = 'contrib/testsuite-management'
- # Pattern for naming manifest files.
- # The first argument should be the toplevel GCC(/GNU tool) source directory.
- # The second argument is the manifest subdir.
- # The third argument is the manifest target, which defaults to the target
- # triplet used during the build.
- _MANIFEST_PATH_PATTERN = '%s/%s/%s.xfail'
- # The options passed to the program.
- _OPTIONS = None
- def Error(msg):
- print >>sys.stderr, 'error: %s' % msg
- sys.exit(1)
- class TestResult(object):
- """Describes a single DejaGNU test result as emitted in .sum files.
- We are only interested in representing unsuccessful tests. So, only
- a subset of all the tests are loaded.
- The summary line used to build the test result should have this format:
- attrlist | XPASS: gcc.dg/unroll_1.c (test for excess errors)
- ^^^^^^^^ ^^^^^ ^^^^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^
- optional state name description
- attributes
- Attributes:
- attrlist: A comma separated list of attributes.
- Valid values:
- flaky Indicates that this test may not always fail. These
- tests are reported, but their presence does not affect
- the results.
- expire=YYYYMMDD After this date, this test will produce an error
- whether it is in the manifest or not.
- state: One of UNRESOLVED, XPASS or FAIL.
- name: File name for the test.
- description: String describing the test (flags used, dejagnu message, etc)
- ordinal: Monotonically increasing integer.
- It is used to keep results for one .exp file sorted
- by the order the tests were run.
- """
- def __init__(self, summary_line, ordinal=-1):
- try:
- (self.attrs, summary_line) = SplitAttributesFromSummaryLine(summary_line)
- try:
- (self.state,
- self.name,
- self.description) = re.match(r'([A-Z]+):\s*(\S+)\s*(.*)',
- summary_line).groups()
- except:
- print 'Failed to parse summary line: "%s"' % summary_line
- raise
- self.ordinal = ordinal
- except ValueError:
- Error('Cannot parse summary line "%s"' % summary_line)
- if self.state not in _VALID_TEST_RESULTS:
- Error('Invalid test result %s in "%s" (parsed as "%s")' % (
- self.state, summary_line, self))
- def __lt__(self, other):
- return (self.name < other.name or
- (self.name == other.name and self.ordinal < other.ordinal))
- def __hash__(self):
- return hash(self.state) ^ hash(self.name) ^ hash(self.description)
- def __eq__(self, other):
- return (self.state == other.state and
- self.name == other.name and
- self.description == other.description)
- def __ne__(self, other):
- return not (self == other)
- def __str__(self):
- attrs = ''
- if self.attrs:
- attrs = '%s | ' % self.attrs
- return '%s%s: %s %s' % (attrs, self.state, self.name, self.description)
- def ExpirationDate(self):
- # Return a datetime.date object with the expiration date for this
- # test result. Return None, if no expiration has been set.
- if re.search(r'expire=', self.attrs):
- expiration = re.search(r'expire=(\d\d\d\d)(\d\d)(\d\d)', self.attrs)
- if not expiration:
- Error('Invalid expire= format in "%s". Must be of the form '
- '"expire=YYYYMMDD"' % self)
- return datetime.date(int(expiration.group(1)),
- int(expiration.group(2)),
- int(expiration.group(3)))
- return None
- def HasExpired(self):
- # Return True if the expiration date of this result has passed.
- expiration_date = self.ExpirationDate()
- if expiration_date:
- now = datetime.date.today()
- return now > expiration_date
- def GetMakefileValue(makefile_name, value_name):
- if os.path.exists(makefile_name):
- makefile = open(makefile_name)
- for line in makefile:
- if line.startswith(value_name):
- (_, value) = line.split('=', 1)
- value = value.strip()
- makefile.close()
- return value
- makefile.close()
- return None
- def ValidBuildDirectory(builddir):
- if (not os.path.exists(builddir) or
- not os.path.exists('%s/Makefile' % builddir)):
- return False
- return True
- def IsComment(line):
- """Return True if line is a comment."""
- return line.startswith('#')
- def SplitAttributesFromSummaryLine(line):
- """Splits off attributes from a summary line, if present."""
- if '|' in line and not _VALID_TEST_RESULTS_REX.match(line):
- (attrs, line) = line.split('|', 1)
- attrs = attrs.strip()
- else:
- attrs = ''
- line = line.strip()
- return (attrs, line)
- def IsInterestingResult(line):
- """Return True if line is one of the summary lines we care about."""
- (_, line) = SplitAttributesFromSummaryLine(line)
- return bool(_VALID_TEST_RESULTS_REX.match(line))
- def IsInclude(line):
- """Return True if line is an include of another file."""
- return line.startswith("@include ")
- def GetIncludeFile(line, includer):
- """Extract the name of the include file from line."""
- includer_dir = os.path.dirname(includer)
- include_file = line[len("@include "):]
- return os.path.join(includer_dir, include_file.strip())
- def IsNegativeResult(line):
- """Return True if line should be removed from the expected results."""
- return line.startswith("@remove ")
- def GetNegativeResult(line):
- """Extract the name of the negative result from line."""
- line = line[len("@remove "):]
- return line.strip()
- def ParseManifestWorker(result_set, manifest_path):
- """Read manifest_path, adding the contents to result_set."""
- if _OPTIONS.verbosity >= 1:
- print 'Parsing manifest file %s.' % manifest_path
- manifest_file = open(manifest_path)
- for line in manifest_file:
- line = line.strip()
- if line == "":
- pass
- elif IsComment(line):
- pass
- elif IsNegativeResult(line):
- result_set.remove(TestResult(GetNegativeResult(line)))
- elif IsInclude(line):
- ParseManifestWorker(result_set, GetIncludeFile(line, manifest_path))
- elif IsInterestingResult(line):
- result_set.add(TestResult(line))
- else:
- Error('Unrecognized line in manifest file: %s' % line)
- manifest_file.close()
- def ParseManifest(manifest_path):
- """Create a set of TestResult instances from the given manifest file."""
- result_set = set()
- ParseManifestWorker(result_set, manifest_path)
- return result_set
- def ParseSummary(sum_fname):
- """Create a set of TestResult instances from the given summary file."""
- result_set = set()
- # ordinal is used when sorting the results so that tests within each
- # .exp file are kept sorted.
- ordinal=0
- sum_file = open(sum_fname)
- for line in sum_file:
- if IsInterestingResult(line):
- result = TestResult(line, ordinal)
- ordinal += 1
- if result.HasExpired():
- # Tests that have expired are not added to the set of expected
- # results. If they are still present in the set of actual results,
- # they will cause an error to be reported.
- print 'WARNING: Expected failure "%s" has expired.' % line.strip()
- continue
- result_set.add(result)
- sum_file.close()
- return result_set
- def GetManifest(manifest_path):
- """Build a set of expected failures from the manifest file.
- Each entry in the manifest file should have the format understood
- by the TestResult constructor.
- If no manifest file exists for this target, it returns an empty set.
- """
- if os.path.exists(manifest_path):
- return ParseManifest(manifest_path)
- else:
- return set()
- def CollectSumFiles(builddir):
- sum_files = []
- for root, dirs, files in os.walk(builddir):
- for ignored in ('.svn', '.git'):
- if ignored in dirs:
- dirs.remove(ignored)
- for fname in files:
- if fname.endswith('.sum'):
- sum_files.append(os.path.join(root, fname))
- return sum_files
- def GetResults(sum_files):
- """Collect all the test results from the given .sum files."""
- build_results = set()
- for sum_fname in sum_files:
- print '\t%s' % sum_fname
- build_results |= ParseSummary(sum_fname)
- return build_results
- def CompareResults(manifest, actual):
- """Compare sets of results and return two lists:
- - List of results present in ACTUAL but missing from MANIFEST.
- - List of results present in MANIFEST but missing from ACTUAL.
- """
- # Collect all the actual results not present in the manifest.
- # Results in this set will be reported as errors.
- actual_vs_manifest = set()
- for actual_result in actual:
- if actual_result not in manifest:
- actual_vs_manifest.add(actual_result)
- # Collect all the tests in the manifest that were not found
- # in the actual results.
- # Results in this set will be reported as warnings (since
- # they are expected failures that are not failing anymore).
- manifest_vs_actual = set()
- for expected_result in manifest:
- # Ignore tests marked flaky.
- if 'flaky' in expected_result.attrs:
- continue
- if expected_result not in actual:
- manifest_vs_actual.add(expected_result)
- return actual_vs_manifest, manifest_vs_actual
- def GetManifestPath(srcdir, target, user_provided_must_exist):
- """Return the full path to the manifest file."""
- manifest_path = _OPTIONS.manifest
- if manifest_path:
- if user_provided_must_exist and not os.path.exists(manifest_path):
- Error('Manifest does not exist: %s' % manifest_path)
- return manifest_path
- else:
- if not srcdir:
- Error('Could not determine the location of GCC\'s source tree. '
- 'The Makefile does not contain a definition for "srcdir".')
- if not target:
- Error('Could not determine the target triplet for this build. '
- 'The Makefile does not contain a definition for "target_alias".')
- return _MANIFEST_PATH_PATTERN % (srcdir, _MANIFEST_SUBDIR, target)
- def GetBuildData():
- if not ValidBuildDirectory(_OPTIONS.build_dir):
- # If we have been given a set of results to use, we may
- # not be inside a valid GCC build directory. In that case,
- # the user must provide both a manifest file and a set
- # of results to check against it.
- if not _OPTIONS.results or not _OPTIONS.manifest:
- Error('%s is not a valid GCC top level build directory. '
- 'You must use --manifest and --results to do the validation.' %
- _OPTIONS.build_dir)
- else:
- return None, None
- srcdir = GetMakefileValue('%s/Makefile' % _OPTIONS.build_dir, 'srcdir =')
- target = GetMakefileValue('%s/Makefile' % _OPTIONS.build_dir, 'target_alias=')
- print 'Source directory: %s' % srcdir
- print 'Build target: %s' % target
- return srcdir, target
- def PrintSummary(msg, summary):
- print '\n\n%s' % msg
- for result in sorted(summary):
- print result
- def GetSumFiles(results, build_dir):
- if not results:
- print 'Getting actual results from build directory %s' % build_dir
- sum_files = CollectSumFiles(build_dir)
- else:
- print 'Getting actual results from user-provided results'
- sum_files = results.split()
- return sum_files
- def PerformComparison(expected, actual, ignore_missing_failures):
- actual_vs_expected, expected_vs_actual = CompareResults(expected, actual)
- tests_ok = True
- if len(actual_vs_expected) > 0:
- PrintSummary('Unexpected results in this build (new failures)',
- actual_vs_expected)
- tests_ok = False
- if not ignore_missing_failures and len(expected_vs_actual) > 0:
- PrintSummary('Expected results not present in this build (fixed tests)'
- '\n\nNOTE: This is not a failure. It just means that these '
- 'tests were expected\nto fail, but either they worked in '
- 'this configuration or they were not\npresent at all.\n',
- expected_vs_actual)
- if tests_ok:
- print '\nSUCCESS: No unexpected failures.'
- return tests_ok
- def CheckExpectedResults():
- srcdir, target = GetBuildData()
- manifest_path = GetManifestPath(srcdir, target, True)
- print 'Manifest: %s' % manifest_path
- manifest = GetManifest(manifest_path)
- sum_files = GetSumFiles(_OPTIONS.results, _OPTIONS.build_dir)
- actual = GetResults(sum_files)
- if _OPTIONS.verbosity >= 1:
- PrintSummary('Tests expected to fail', manifest)
- PrintSummary('\nActual test results', actual)
- return PerformComparison(manifest, actual, _OPTIONS.ignore_missing_failures)
- def ProduceManifest():
- (srcdir, target) = GetBuildData()
- manifest_path = GetManifestPath(srcdir, target, False)
- print 'Manifest: %s' % manifest_path
- if os.path.exists(manifest_path) and not _OPTIONS.force:
- Error('Manifest file %s already exists.\nUse --force to overwrite.' %
- manifest_path)
- sum_files = GetSumFiles(_OPTIONS.results, _OPTIONS.build_dir)
- actual = GetResults(sum_files)
- manifest_file = open(manifest_path, 'w')
- for result in sorted(actual):
- print result
- manifest_file.write('%s\n' % result)
- manifest_file.close()
- return True
- def CompareBuilds():
- (srcdir, target) = GetBuildData()
- sum_files = GetSumFiles(_OPTIONS.results, _OPTIONS.build_dir)
- actual = GetResults(sum_files)
- clean_sum_files = GetSumFiles(_OPTIONS.results, _OPTIONS.clean_build)
- clean = GetResults(clean_sum_files)
- return PerformComparison(clean, actual, _OPTIONS.ignore_missing_failures)
- def Main(argv):
- parser = optparse.OptionParser(usage=__doc__)
- # Keep the following list sorted by option name.
- parser.add_option('--build_dir', action='store', type='string',
- dest='build_dir', default='.',
- help='Build directory to check (default = .)')
- parser.add_option('--clean_build', action='store', type='string',
- dest='clean_build', default=None,
- help='Compare test results from this build against '
- 'those of another (clean) build. Use this option '
- 'when comparing the test results of your patch versus '
- 'the test results of a clean build without your patch. '
- 'You must provide the path to the top directory of your '
- 'clean build.')
- parser.add_option('--force', action='store_true', dest='force',
- default=False, help='When used with --produce_manifest, '
- 'it will overwrite an existing manifest file '
- '(default = False)')
- parser.add_option('--ignore_missing_failures', action='store_true',
- dest='ignore_missing_failures', default=False,
- help='When a failure is expected in the manifest but '
- 'it is not found in the actual results, the script '
- 'produces a note alerting to this fact. This means '
- 'that the expected failure has been fixed, or '
- 'it did not run, or it may simply be flaky '
- '(default = False)')
- parser.add_option('--manifest', action='store', type='string',
- dest='manifest', default=None,
- help='Name of the manifest file to use (default = '
- 'taken from '
- 'contrib/testsuite-managment/<target_alias>.xfail)')
- parser.add_option('--produce_manifest', action='store_true',
- dest='produce_manifest', default=False,
- help='Produce the manifest for the current '
- 'build (default = False)')
- parser.add_option('--results', action='store', type='string',
- dest='results', default=None, help='Space-separated list '
- 'of .sum files with the testing results to check. The '
- 'only content needed from these files are the lines '
- 'starting with FAIL, XPASS or UNRESOLVED (default = '
- '.sum files collected from the build directory).')
- parser.add_option('--verbosity', action='store', dest='verbosity',
- type='int', default=0, help='Verbosity level (default = 0)')
- global _OPTIONS
- (_OPTIONS, _) = parser.parse_args(argv[1:])
- if _OPTIONS.produce_manifest:
- retval = ProduceManifest()
- elif _OPTIONS.clean_build:
- retval = CompareBuilds()
- else:
- retval = CheckExpectedResults()
- if retval:
- return 0
- else:
- return 1
- if __name__ == '__main__':
- retval = Main(sys.argv)
- sys.exit(retval)
|