1 #
   2 # ----------------------------------------------------------------------------------------------------
   3 #
   4 # Copyright (c) 2007, 2015, Oracle and/or its affiliates. All rights reserved.
   5 # DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   6 #
   7 # This code is free software; you can redistribute it and/or modify it
   8 # under the terms of the GNU General Public License version 2 only, as
   9 # published by the Free Software Foundation.
  10 #
  11 # This code is distributed in the hope that it will be useful, but WITHOUT
  12 # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  13 # FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  14 # version 2 for more details (a copy is included in the LICENSE file that
  15 # accompanied this code).
  16 #
  17 # You should have received a copy of the GNU General Public License version
  18 # 2 along with this work; if not, write to the Free Software Foundation,
  19 # Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  20 #
  21 # Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  22 # or visit www.oracle.com if you need additional information or have any
  23 # questions.
  24 #
  25 # ----------------------------------------------------------------------------------------------------
  26 
  27 import sanitycheck
  28 import itertools
  29 import json
  30 
  31 import mx
  32 import mx_graal
  33 
  34 def _run_benchmark(args, availableBenchmarks, runBenchmark):
  35 
  36     vmOpts, benchmarksAndOptions = mx.extract_VM_args(args, useDoubleDash=availableBenchmarks is None)
  37 
  38     if availableBenchmarks is None:
  39         harnessArgs = benchmarksAndOptions
  40         return runBenchmark(None, harnessArgs, vmOpts)
  41 
  42     if len(benchmarksAndOptions) == 0:
  43         mx.abort('at least one benchmark name or "all" must be specified')
  44     benchmarks = list(itertools.takewhile(lambda x: not x.startswith('-'), benchmarksAndOptions))
  45     harnessArgs = benchmarksAndOptions[len(benchmarks):]
  46 
  47     if 'all' in benchmarks:
  48         benchmarks = availableBenchmarks
  49     else:
  50         for bm in benchmarks:
  51             if bm not in availableBenchmarks:
  52                 mx.abort('unknown benchmark: ' + bm + '\nselect one of: ' + str(availableBenchmarks))
  53 
  54     failed = []
  55     for bm in benchmarks:
  56         if not runBenchmark(bm, harnessArgs, vmOpts):
  57             failed.append(bm)
  58 
  59     if len(failed) != 0:
  60         mx.abort('Benchmark failures: ' + str(failed))
  61 
  62 def deoptalot(args):
  63     """bootstrap a VM with DeoptimizeALot and VerifyOops on
  64 
  65     If the first argument is a number, the process will be repeated
  66     this number of times. All other arguments are passed to the VM."""
  67     count = 1
  68     if len(args) > 0 and args[0].isdigit():
  69         count = int(args[0])
  70         del args[0]
  71 
  72     for _ in range(count):
  73         if not mx_graal.run_vm(['-XX:-TieredCompilation', '-XX:+DeoptimizeALot', '-XX:+VerifyOops'] + args + ['-version']) == 0:
  74             mx.abort("Failed")
  75 
  76 def longtests(args):
  77 
  78     deoptalot(['15', '-Xmx48m'])
  79 
  80     dacapo(['100', 'eclipse', '-esa'])
  81 
  82 def dacapo(args):
  83     """run one or more DaCapo benchmarks"""
  84 
  85     def launcher(bm, harnessArgs, extraVmOpts):
  86         return sanitycheck.getDacapo(bm, harnessArgs).test(mx_graal.get_vm(), extraVmOpts=extraVmOpts)
  87 
  88     _run_benchmark(args, sanitycheck.dacapoSanityWarmup.keys(), launcher)
  89 
  90 def scaladacapo(args):
  91     """run one or more Scala DaCapo benchmarks"""
  92 
  93     def launcher(bm, harnessArgs, extraVmOpts):
  94         return sanitycheck.getScalaDacapo(bm, harnessArgs).test(mx_graal.get_vm(), extraVmOpts=extraVmOpts)
  95 
  96     _run_benchmark(args, sanitycheck.dacapoScalaSanityWarmup.keys(), launcher)
  97 
  98 
  99 """
 100 Extra benchmarks to run from 'bench()'.
 101 """
 102 extraBenchmarks = []
 103 
 104 def bench(args):
 105     """run benchmarks and parse their output for results
 106 
 107     Results are JSON formated : {group : {benchmark : score}}."""
 108     resultFile = None
 109     if '-resultfile' in args:
 110         index = args.index('-resultfile')
 111         if index + 1 < len(args):
 112             resultFile = args[index + 1]
 113             del args[index]
 114             del args[index]
 115         else:
 116             mx.abort('-resultfile must be followed by a file name')
 117     resultFileCSV = None
 118     if '-resultfilecsv' in args:
 119         index = args.index('-resultfilecsv')
 120         if index + 1 < len(args):
 121             resultFileCSV = args[index + 1]
 122             del args[index]
 123             del args[index]
 124         else:
 125             mx.abort('-resultfilecsv must be followed by a file name')
 126     vm = mx_graal.get_vm()
 127     if len(args) is 0:
 128         args = ['all']
 129 
 130     vmArgs = [arg for arg in args if arg.startswith('-')]
 131 
 132     def benchmarks_in_group(group):
 133         prefix = group + ':'
 134         return [a[len(prefix):] for a in args if a.startswith(prefix)]
 135 
 136     results = {}
 137     benchmarks = []
 138     # DaCapo
 139     if 'dacapo' in args or 'all' in args:
 140         benchmarks += sanitycheck.getDacapos(level=sanitycheck.SanityCheckLevel.Benchmark)
 141     else:
 142         dacapos = benchmarks_in_group('dacapo')
 143         for dacapo in dacapos:
 144             if dacapo not in sanitycheck.dacapoSanityWarmup.keys():
 145                 mx.abort('Unknown DaCapo : ' + dacapo)
 146             iterations = sanitycheck.dacapoSanityWarmup[dacapo][sanitycheck.SanityCheckLevel.Benchmark]
 147             if iterations > 0:
 148                 benchmarks += [sanitycheck.getDacapo(dacapo, ['-n', str(iterations)])]
 149 
 150     if 'scaladacapo' in args or 'all' in args:
 151         benchmarks += sanitycheck.getScalaDacapos(level=sanitycheck.SanityCheckLevel.Benchmark)
 152     else:
 153         scaladacapos = benchmarks_in_group('scaladacapo')
 154         for scaladacapo in scaladacapos:
 155             if scaladacapo not in sanitycheck.dacapoScalaSanityWarmup.keys():
 156                 mx.abort('Unknown Scala DaCapo : ' + scaladacapo)
 157             iterations = sanitycheck.dacapoScalaSanityWarmup[scaladacapo][sanitycheck.SanityCheckLevel.Benchmark]
 158             if iterations > 0:
 159                 benchmarks += [sanitycheck.getScalaDacapo(scaladacapo, ['-n', str(iterations)])]
 160 
 161     # Bootstrap
 162     if 'bootstrap' in args or 'all' in args:
 163         benchmarks += sanitycheck.getBootstraps()
 164     # SPECjvm2008
 165     if 'specjvm2008' in args or 'all' in args:
 166         benchmarks += [sanitycheck.getSPECjvm2008(['-ikv', '-wt', '120', '-it', '120'])]
 167     else:
 168         specjvms = benchmarks_in_group('specjvm2008')
 169         for specjvm in specjvms:
 170             benchmarks += [sanitycheck.getSPECjvm2008(['-ikv', '-wt', '120', '-it', '120', specjvm])]
 171 
 172     if 'specjbb2005' in args or 'all' in args:
 173         benchmarks += [sanitycheck.getSPECjbb2005()]
 174 
 175     if 'specjbb2013' in args:  # or 'all' in args //currently not in default set
 176         benchmarks += [sanitycheck.getSPECjbb2013()]
 177 
 178     if 'ctw-full' in args:
 179         benchmarks.append(sanitycheck.getCTW(vm, sanitycheck.CTWMode.Full))
 180     if 'ctw-noinline' in args:
 181         benchmarks.append(sanitycheck.getCTW(vm, sanitycheck.CTWMode.NoInline))
 182 
 183     for f in extraBenchmarks:
 184         f(args, vm, benchmarks)
 185 
 186     for test in benchmarks:
 187         for (groupName, res) in test.bench(vm, extraVmOpts=vmArgs).items():
 188             group = results.setdefault(groupName, {})
 189             group.update(res)
 190     mx.log(json.dumps(results))
 191     if resultFile:
 192         with open(resultFile, 'w') as f:
 193             f.write(json.dumps(results))
 194     if resultFileCSV:
 195         with open(resultFileCSV, 'w') as f:
 196             for key1, value1 in results.iteritems():
 197                 f.write('%s;\n' % (str(key1)))
 198                 for key2, value2 in sorted(value1.iteritems()):
 199                     f.write('%s; %s;\n' % (str(key2), str(value2)))
 200 
 201 def specjvm2008(args):
 202     """run one or more SPECjvm2008 benchmarks"""
 203 
 204     def launcher(bm, harnessArgs, extraVmOpts):
 205         return sanitycheck.getSPECjvm2008(harnessArgs + [bm]).bench(mx_graal.get_vm(), extraVmOpts=extraVmOpts)
 206 
 207     availableBenchmarks = set(sanitycheck.specjvm2008Names)
 208     if "all" not in args:
 209         # only add benchmark groups if we are not running "all"
 210         for name in sanitycheck.specjvm2008Names:
 211             parts = name.rsplit('.', 1)
 212             if len(parts) > 1:
 213                 assert len(parts) == 2
 214                 group = parts[0]
 215                 availableBenchmarks.add(group)
 216 
 217     _run_benchmark(args, sorted(availableBenchmarks), launcher)
 218 
 219 def specjbb2013(args):
 220     """run the composite SPECjbb2013 benchmark"""
 221 
 222     def launcher(bm, harnessArgs, extraVmOpts):
 223         assert bm is None
 224         return sanitycheck.getSPECjbb2013(harnessArgs).bench(mx_graal.get_vm(), extraVmOpts=extraVmOpts)
 225 
 226     _run_benchmark(args, None, launcher)
 227 
 228 def specjbb2015(args):
 229     """run the composite SPECjbb2015 benchmark"""
 230 
 231     def launcher(bm, harnessArgs, extraVmOpts):
 232         assert bm is None
 233         return sanitycheck.getSPECjbb2015(harnessArgs).bench(mx_graal.get_vm(), extraVmOpts=extraVmOpts)
 234 
 235     _run_benchmark(args, None, launcher)
 236 
 237 def specjbb2005(args):
 238     """run the composite SPECjbb2005 benchmark"""
 239 
 240     def launcher(bm, harnessArgs, extraVmOpts):
 241         assert bm is None
 242         return sanitycheck.getSPECjbb2005(harnessArgs).bench(mx_graal.get_vm(), extraVmOpts=extraVmOpts)
 243 
 244     _run_benchmark(args, None, launcher)
 245 
 246 mx.update_commands(mx.suite('graal'), {
 247     'dacapo': [dacapo, '[VM options] benchmarks...|"all" [DaCapo options]'],
 248     'scaladacapo': [scaladacapo, '[VM options] benchmarks...|"all" [Scala DaCapo options]'],
 249     'specjvm2008': [specjvm2008, '[VM options] benchmarks...|"all" [SPECjvm2008 options]'],
 250     'specjbb2013': [specjbb2013, '[VM options] [-- [SPECjbb2013 options]]'],
 251     'specjbb2015': [specjbb2015, '[VM options] [-- [SPECjbb2015 options]]'],
 252     'specjbb2005': [specjbb2005, '[VM options] [-- [SPECjbb2005 options]]'],
 253     'bench' : [bench, '[-resultfile file] [all(default)|dacapo|specjvm2008|bootstrap]'],
 254     'deoptalot' : [deoptalot, '[n]'],
 255     'longtests' : [longtests, ''],
 256 })