#!/usr/bin/env python from __future__ import print_function import glob, os, re import optparse import inspect """ Quick script for parsing the output of the test system and summarizing the results. """ def inInstallDir(): """ When petsc is installed then this file in installed in: /share/petsc/examples/config/gmakegentest.py otherwise the path is: /config/gmakegentest.py We use this difference to determine if we are in installdir """ thisscriptdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))) dirlist=thisscriptdir.split(os.path.sep) if len(dirlist)>4: lastfour=os.path.sep.join(dirlist[len(dirlist)-4:]) if lastfour==os.path.join('share','petsc','examples','config'): return True else: return False else: return False def summarize_results(directory,make,ntime,etime): ''' Loop over all of the results files and summarize the results''' startdir = os.getcwd() try: os.chdir(directory) except OSError: print('# No tests run') return summary={'total':0,'success':0,'failed':0,'failures':[],'todo':0,'skip':0, 'time':0, 'cputime':0} timesummary={} cputimesummary={} timelist=[] for cfile in glob.glob('*.counts'): with open(cfile, 'r') as f: for line in f: l = line.split() if l[0] == 'failures': if len(l)>1: summary[l[0]] += l[1:] elif l[0] == 'time': if len(l)==1: continue summary[l[0]] += float(l[1]) summary['cputime'] += float(l[2]) timesummary[cfile]=float(l[1]) cputimesummary[cfile]=float(l[2]) timelist.append(float(l[1])) elif l[0] not in summary: continue else: summary[l[0]] += int(l[1]) failstr=' '.join(summary['failures']) print("\n# -------------") print("# Summary ") print("# -------------") if failstr.strip(): print("# FAILED " + failstr) for t in "success failed todo skip".split(): percent=summary[t]/float(summary['total'])*100 print("# %s %d/%d tests (%3.1f%%)" % (t, summary[t], summary['total'], percent)) print("#") if etime: print("# Wall clock time for tests: %s sec"% etime) print("# Approximate CPU time (not incl. build time): %s sec"% summary['cputime']) if failstr.strip(): fail_targets=( re.sub('(?<=[0-9]_\w)_.*','', re.sub('cmd-','', re.sub('diff-','',failstr+' '))) ) # Strip off characters from subtests fail_list=[] for failure in fail_targets.split(): if failure.count('-')>1: fail_list.append('-'.join(failure.split('-')[:-1])) else: fail_list.append(failure) fail_list=list(set(fail_list)) fail_targets=' '.join(fail_list) #Make the message nice makefile="gmakefile.test" if inInstallDir() else "gmakefile" print("#\n# To rerun failed tests: ") print("# "+make+" -f "+makefile+" test search='" + fail_targets.strip()+"'") if ntime>0: print("#\n# Timing summary (actual test time / total CPU time): ") timelist=list(set(timelist)) timelist.sort(reverse=True) nlim=(ntime if ntime0: testdata[pkgname]['errors'] += 1 with open(cfile, 'r') as f: for line in f: l = line.split() if l[0] == 'failed': testdata[pkgname]['problems'][probname][l[0]] = True testdata[pkgname][l[0]] += 1 elif l[0] == 'time': if len(l)==1: continue testdata[pkgname]['problems'][probname][l[0]] = float(l[1]) testdata[pkgname][l[0]] += float(l[1]) elif l[0] == 'skip': testdata[pkgname]['problems'][probname][l[0]] = True testdata[pkgname][l[0]] += 1 elif l[0] not in testdata[pkgname].keys(): continue else: testdata[pkgname][l[0]] += 1 # at this point we have the complete test results in dictionary structures # we can now write this information into a jUnit formatted XLM file junit = open('../testresults.xml', 'w') junit.write('\n') junit.write('\n') for pkg in testdata.keys(): testsuite = testdata[pkg] junit.write(' \n'%( testsuite['errors'], testsuite['failed'], pkg, testsuite['total'])) for prob in testsuite['problems'].keys(): p = testsuite['problems'][prob] junit.write(' \n'%( p['classname'], prob, p['time'])) if p['skipped']: # if we got here, the TAP output shows a skipped test junit.write(' \n') elif len(p['stderr'])>0: # if we got here, the test crashed with an error # we show the stderr output under junit.write(' \n') junit.write("") junit.write(' \n') elif len(p['diff'])>0: # if we got here, the test output did not match the stored output file # we show the diff between new output and old output under junit.write(' \n') junit.write("") junit.write(' \n') elif len(p['stdout'])>0: # if we got here, the test succeeded so we just show the stdout # for manual sanity-checks junit.write(' \n') junit.write("= 1024: break junit.write("]]>") junit.write(' \n') junit.write(' \n') junit.write(' \n') junit.write('') junit.close() os.chdir(startdir) return def main(): parser = optparse.OptionParser(usage="%prog [options]") parser.add_option('-d', '--directory', dest='directory', help='Directory containing results of petsc test system', default=os.path.join(os.environ.get('PETSC_ARCH',''), 'tests','counts')) parser.add_option('-e', '--elapsed_time', dest='elapsed_time', help='Report elapsed time in output', default=None) parser.add_option('-m', '--make', dest='make', help='make executable to report in summary', default='make') parser.add_option('-t', '--time', dest='time', help='-t n: Report on the n number expensive jobs', default=0) options, args = parser.parse_args() # Process arguments if len(args) > 0: parser.print_usage() return summarize_results(options.directory,options.make,int(options.time),options.elapsed_time) generate_xml(options.directory) if __name__ == "__main__": main()