#!/usr/bin/env python from __future__ import print_function import glob, os, re, stat import optparse import inspect """ Quick script for parsing the output of the test system and summarizing the results. """ def inInstallDir(): """ When petsc is installed then this file in installed in: /share/petsc/examples/config/gmakegentest.py otherwise the path is: /config/gmakegentest.py We use this difference to determine if we are in installdir """ thisscriptdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))) dirlist=thisscriptdir.split(os.path.sep) if len(dirlist)>4: lastfour=os.path.sep.join(dirlist[len(dirlist)-4:]) if lastfour==os.path.join('share','petsc','examples','config'): return True else: return False else: return False def summarize_results(directory,make,ntime,etime): ''' Loop over all of the results files and summarize the results''' startdir = os.getcwd() try: os.chdir(directory) except OSError: print('# No tests run') return summary={'total':0,'success':0,'failed':0,'failures':[],'todo':0,'skip':0, 'time':0, 'cputime':0} timesummary={} cputimesummary={} timelist=[] for cfile in glob.glob('*.counts'): with open(cfile, 'r') as f: for line in f: l = line.split() if l[0] == 'failures': if len(l)>1: summary[l[0]] += l[1:] elif l[0] == 'time': if len(l)==1: continue summary[l[0]] += float(l[1]) summary['cputime'] += float(l[2]) timesummary[cfile]=float(l[1]) cputimesummary[cfile]=float(l[2]) timelist.append(float(l[1])) elif l[0] not in summary: continue else: summary[l[0]] += int(l[1]) failstr=' '.join(summary['failures']) print("\n# -------------") print("# Summary ") print("# -------------") if failstr.strip(): print("# FAILED " + failstr) for t in "success failed todo skip".split(): percent=summary[t]/float(summary['total'])*100 print("# %s %d/%d tests (%3.1f%%)" % (t, summary[t], summary['total'], percent)) print("#") if etime: print("# Wall clock time for tests: %s sec"% etime) print("# Approximate CPU time (not incl. build time): %s sec"% summary['cputime']) if failstr.strip(): fail_targets=( re.sub('cmd-','', re.sub('diff-','',failstr+' ')) ) # Strip off characters from subtests fail_list=[] for failure in fail_targets.split(): fail_list.append(failure.split('+')[0]) fail_list=list(set(fail_list)) fail_targets=' '.join(fail_list) # create simple little script sfile=os.path.join(os.path.dirname(os.path.abspath(os.curdir)),'echofailures.sh') with open(sfile,'w') as f: f.write('echo '+fail_targets.strip()) st = os.stat(sfile) os.chmod(sfile, st.st_mode | stat.S_IEXEC) #Make the message nice makefile="gmakefile.test" if inInstallDir() else "gmakefile" print("#\n# To rerun failed tests: ") print("# "+make+" -f "+makefile+" test test-fail=1") if ntime>0: print("#\n# Timing summary (actual test time / total CPU time): ") timelist=list(set(timelist)) timelist.sort(reverse=True) nlim=(ntime if ntime0: testdata[pkgname]['errors'] += 1 with open(cfile, 'r') as f: for line in f: l = line.split() if l[0] == 'time': if len(l)==1: continue testdata[pkgname]['problems'][probname][l[0]] = float(l[1]) testdata[pkgname][l[0]] += float(l[1]) elif l[0] in testdata[pkgname].keys(): num_int=int(l[1]) testdata[pkgname][l[0]] += num_int if l[0] in ['failed','skip'] and num_int: testdata[pkgname]['problems'][probname][l[0]] = True else: continue os.chdir(startdir) # Keep function in good state return testdata def show_fail(testdata): """ Show the failures and commands to run them """ for pkg in testdata.keys(): testsuite = testdata[pkg] for prob in testsuite['problems'].keys(): p = testsuite['problems'][prob] cdbase='cd '+p['probdir']+' && ' if p['skipped']: # if we got here, the TAP output shows a skipped test pass elif len(p['stderr'])>0: # if we got here, the test crashed with an error # we show the stderr output under shbase=os.path.join(p['probdir'], p['fullname']) shfile=shbase+".sh" if not os.path.exists(shfile): shfile=glob.glob(shbase+"*")[0] with open(shfile, 'r') as sh: cmd = sh.read() print(p['fullname']+': '+cdbase+cmd.split('>')[0]) elif len(p['diff'])>0: # if we got here, the test output did not match the stored output file # we show the diff between new output and old output under shbase=os.path.join(p['probdir'], 'diff-'+p['fullname']) shfile=shbase+".sh" if not os.path.exists(shfile): shfile=glob.glob(shbase+"*")[0] with open(shfile, 'r') as sh: cmd = sh.read() print(p['fullname']+': '+cdbase+cmd.split('>')[0]) pass return def generate_xml(testdata,directory): """ write testdata information into a jUnit formatted XLM file """ startdir= os.getcwd() try: os.chdir(directory) except OSError: print('# No tests run') return junit = open('../testresults.xml', 'w') junit.write('\n') junit.write('\n') for pkg in testdata.keys(): testsuite = testdata[pkg] junit.write(' \n'%( testsuite['errors'], testsuite['failed'], pkg, testsuite['total'])) for prob in testsuite['problems'].keys(): p = testsuite['problems'][prob] junit.write(' \n'%( p['classname'], prob, p['time'])) if p['skipped']: # if we got here, the TAP output shows a skipped test junit.write(' \n') elif len(p['stderr'])>0: # if we got here, the test crashed with an error # we show the stderr output under junit.write(' \n') junit.write("0: for line in p['stdout']: junit.write("%s\n"%line.rstrip()) for line in p['stderr']: junit.write("%s\n"%line.rstrip()) junit.write("]]>") junit.write(' \n') elif len(p['diff'])>0: # if we got here, the test output did not match the stored output file # we show the diff between new output and old output under junit.write(' \n') junit.write("") junit.write(' \n') junit.write(' \n') junit.write(' \n') junit.write('') junit.close() os.chdir(startdir) return def main(): parser = optparse.OptionParser(usage="%prog [options]") parser.add_option('-d', '--directory', dest='directory', help='Directory containing results of petsc test system', default=os.path.join(os.environ.get('PETSC_ARCH',''), 'tests','counts')) parser.add_option('-e', '--elapsed_time', dest='elapsed_time', help='Report elapsed time in output', default=None) parser.add_option('-m', '--make', dest='make', help='make executable to report in summary', default='make') parser.add_option('-t', '--time', dest='time', help='-t n: Report on the n number expensive jobs', default=0) parser.add_option('-f', '--fail', dest='show_fail', action="store_true", help='Show the failed tests and how to run them') options, args = parser.parse_args() # Process arguments if len(args) > 0: parser.print_usage() return if not options.show_fail: summarize_results(options.directory,options.make,int(options.time),options.elapsed_time) testresults=get_test_data(options.directory) if options.show_fail: show_fail(testresults) else: generate_xml(testresults, options.directory) if __name__ == "__main__": main()