43972
|
1 |
# ----------------------------------------------------------------------------------------------------
|
|
2 |
#
|
|
3 |
# Copyright (c) 2007, 2012, Oracle and/or its affiliates. All rights reserved.
|
|
4 |
# DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
5 |
#
|
|
6 |
# This code is free software; you can redistribute it and/or modify it
|
|
7 |
# under the terms of the GNU General Public License version 2 only, as
|
|
8 |
# published by the Free Software Foundation.
|
|
9 |
#
|
|
10 |
# This code is distributed in the hope that it will be useful, but WITHOUT
|
|
11 |
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
12 |
# FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
13 |
# version 2 for more details (a copy is included in the LICENSE file that
|
|
14 |
# accompanied this code).
|
|
15 |
#
|
|
16 |
# You should have received a copy of the GNU General Public License version
|
|
17 |
# 2 along with this work; if not, write to the Free Software Foundation,
|
|
18 |
# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
19 |
#
|
|
20 |
# Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
21 |
# or visit www.oracle.com if you need additional information or have any
|
|
22 |
# questions.
|
|
23 |
#
|
|
24 |
# ----------------------------------------------------------------------------------------------------
|
|
25 |
|
|
26 |
from outputparser import OutputParser, ValuesMatcher
|
|
27 |
import re, mx, mx_graal, os, sys, StringIO, subprocess
|
|
28 |
from os.path import isfile, join, exists
|
|
29 |
|
|
30 |
gc = 'UseSerialGC'
|
|
31 |
|
|
32 |
dacapoSanityWarmup = {
|
|
33 |
'avrora': [0, 0, 3, 6, 13],
|
|
34 |
'batik': [0, 0, 5, 5, 20],
|
|
35 |
'eclipse': [0, 0, 0, 0, 0],
|
|
36 |
'fop': [4, 8, 10, 20, 30],
|
|
37 |
'h2': [0, 0, 5, 5, 8],
|
|
38 |
'jython': [0, 0, 5, 10, 13],
|
|
39 |
'luindex': [0, 0, 5, 10, 10],
|
|
40 |
'lusearch': [0, 4, 5, 5, 8],
|
|
41 |
'pmd': [0, 0, 5, 10, 13],
|
|
42 |
'sunflow': [0, 2, 5, 10, 15],
|
|
43 |
'tomcat': [0, 0, 5, 10, 15],
|
|
44 |
'tradebeans': [0, 0, 5, 10, 13],
|
|
45 |
'tradesoap': [0, 0, 5, 10, 15],
|
|
46 |
'xalan': [0, 0, 5, 10, 18],
|
|
47 |
}
|
|
48 |
|
|
49 |
dacapoScalaSanityWarmup = {
|
|
50 |
'actors': [0, 0, 2, 5, 5],
|
|
51 |
'apparat': [0, 0, 2, 5, 5],
|
|
52 |
'factorie': [0, 0, 2, 5, 5],
|
|
53 |
'kiama': [0, 4, 3, 13, 15],
|
|
54 |
'scalac': [0, 0, 5, 15, 20],
|
|
55 |
'scaladoc': [0, 0, 5, 15, 15],
|
|
56 |
'scalap': [0, 0, 5, 15, 20],
|
|
57 |
'scalariform':[0, 0, 6, 15, 20],
|
|
58 |
'scalatest': [0, 0, 2, 10, 12],
|
|
59 |
'scalaxb': [0, 0, 5, 15, 25],
|
|
60 |
# (gdub) specs sometimes returns a non-zero value event though there is no apparent failure
|
|
61 |
'specs': [0, 0, 0, 0, 0],
|
|
62 |
'tmt': [0, 0, 3, 10, 12]
|
|
63 |
}
|
|
64 |
|
|
65 |
dacapoGateBuildLevels = {
|
|
66 |
'avrora': ['product', 'fastdebug', 'debug'],
|
|
67 |
'batik': ['product', 'fastdebug', 'debug'],
|
|
68 |
# (lewurm): does not work with JDK8
|
|
69 |
'eclipse': [],
|
|
70 |
'fop': ['fastdebug', 'debug'],
|
|
71 |
'h2': ['product', 'fastdebug', 'debug'],
|
|
72 |
'jython': ['product', 'fastdebug', 'debug'],
|
|
73 |
'luindex': ['product', 'fastdebug', 'debug'],
|
|
74 |
'lusearch': ['product'],
|
|
75 |
'pmd': ['product', 'fastdebug', 'debug'],
|
|
76 |
'sunflow': ['fastdebug', 'debug'],
|
|
77 |
'tomcat': ['product', 'fastdebug', 'debug'],
|
|
78 |
'tradebeans': ['product', 'fastdebug', 'debug'],
|
|
79 |
# tradesoap is too unreliable for the gate, often crashing with concurrency problems:
|
|
80 |
# http://sourceforge.net/p/dacapobench/bugs/99/
|
|
81 |
'tradesoap': [],
|
|
82 |
'xalan': ['product', 'fastdebug', 'debug'],
|
|
83 |
}
|
|
84 |
|
|
85 |
dacapoScalaGateBuildLevels = {
|
|
86 |
'actors': ['product', 'fastdebug', 'debug'],
|
|
87 |
'apparat': ['product', 'fastdebug', 'debug'],
|
|
88 |
'factorie': ['product', 'fastdebug', 'debug'],
|
|
89 |
'kiama': ['fastdebug', 'debug'],
|
|
90 |
'scalac': ['product', 'fastdebug', 'debug'],
|
|
91 |
'scaladoc': ['product', 'fastdebug', 'debug'],
|
|
92 |
'scalap': ['product', 'fastdebug', 'debug'],
|
|
93 |
'scalariform':['product', 'fastdebug', 'debug'],
|
|
94 |
'scalatest': ['product', 'fastdebug', 'debug'],
|
|
95 |
'scalaxb': ['product', 'fastdebug', 'debug'],
|
|
96 |
'specs': ['product', 'fastdebug', 'debug'],
|
|
97 |
'tmt': ['product', 'fastdebug', 'debug'],
|
|
98 |
}
|
|
99 |
|
|
100 |
specjvm2008Names = [
|
|
101 |
'startup.helloworld',
|
|
102 |
'startup.compiler.compiler',
|
|
103 |
'startup.compiler.sunflow',
|
|
104 |
'startup.compress',
|
|
105 |
'startup.crypto.aes',
|
|
106 |
'startup.crypto.rsa',
|
|
107 |
'startup.crypto.signverify',
|
|
108 |
'startup.mpegaudio',
|
|
109 |
'startup.scimark.fft',
|
|
110 |
'startup.scimark.lu',
|
|
111 |
'startup.scimark.monte_carlo',
|
|
112 |
'startup.scimark.sor',
|
|
113 |
'startup.scimark.sparse',
|
|
114 |
'startup.serial',
|
|
115 |
'startup.sunflow',
|
|
116 |
'startup.xml.transform',
|
|
117 |
'startup.xml.validation',
|
|
118 |
'compiler.compiler',
|
|
119 |
'compiler.sunflow',
|
|
120 |
'compress',
|
|
121 |
'crypto.aes',
|
|
122 |
'crypto.rsa',
|
|
123 |
'crypto.signverify',
|
|
124 |
'derby',
|
|
125 |
'mpegaudio',
|
|
126 |
'scimark.fft.large',
|
|
127 |
'scimark.lu.large',
|
|
128 |
'scimark.sor.large',
|
|
129 |
'scimark.sparse.large',
|
|
130 |
'scimark.fft.small',
|
|
131 |
'scimark.lu.small',
|
|
132 |
'scimark.sor.small',
|
|
133 |
'scimark.sparse.small',
|
|
134 |
'scimark.monte_carlo',
|
|
135 |
'serial',
|
|
136 |
'sunflow',
|
|
137 |
'xml.transform',
|
|
138 |
'xml.validation'
|
|
139 |
]
|
|
140 |
|
|
141 |
def _noneAsEmptyList(a):
|
|
142 |
if a is None:
|
|
143 |
return []
|
|
144 |
return a
|
|
145 |
|
|
146 |
class SanityCheckLevel:
|
|
147 |
Fast, Gate, Normal, Extensive, Benchmark = range(5)
|
|
148 |
|
|
149 |
def getSPECjbb2005(benchArgs=None):
|
|
150 |
benchArgs = [] if benchArgs is None else benchArgs
|
|
151 |
|
|
152 |
specjbb2005 = mx.get_env('SPECJBB2005')
|
|
153 |
if specjbb2005 is None or not exists(join(specjbb2005, 'jbb.jar')):
|
|
154 |
mx.abort('Please set the SPECJBB2005 environment variable to a SPECjbb2005 directory')
|
|
155 |
|
|
156 |
score = re.compile(r"^Valid run, Score is (?P<score>[0-9]+)$", re.MULTILINE)
|
|
157 |
error = re.compile(r"VALIDATION ERROR")
|
|
158 |
success = re.compile(r"^Valid run, Score is [0-9]+$", re.MULTILINE)
|
|
159 |
matcher = ValuesMatcher(score, {'group' : 'SPECjbb2005', 'name' : 'score', 'score' : '<score>'})
|
|
160 |
classpath = ['jbb.jar', 'check.jar']
|
|
161 |
return Test("SPECjbb2005", ['spec.jbb.JBBmain', '-propfile', 'SPECjbb.props'] + benchArgs, [success], [error], [matcher], vmOpts=['-Xms3g', '-XX:+' + gc, '-XX:-UseCompressedOops', '-cp', os.pathsep.join(classpath)], defaultCwd=specjbb2005)
|
|
162 |
|
|
163 |
def getSPECjbb2013(benchArgs=None):
|
|
164 |
|
|
165 |
specjbb2013 = mx.get_env('SPECJBB2013')
|
|
166 |
if specjbb2013 is None or not exists(join(specjbb2013, 'specjbb2013.jar')):
|
|
167 |
mx.abort('Please set the SPECJBB2013 environment variable to a SPECjbb2013 directory')
|
|
168 |
|
|
169 |
jops = re.compile(r"^RUN RESULT: hbIR \(max attempted\) = [0-9]+, hbIR \(settled\) = [0-9]+, max-jOPS = (?P<max>[0-9]+), critical-jOPS = (?P<critical>[0-9]+)$", re.MULTILINE)
|
|
170 |
# error?
|
|
171 |
success = re.compile(r"org.spec.jbb.controller: Run finished", re.MULTILINE)
|
|
172 |
matcherMax = ValuesMatcher(jops, {'group' : 'SPECjbb2013', 'name' : 'max', 'score' : '<max>'})
|
|
173 |
matcherCritical = ValuesMatcher(jops, {'group' : 'SPECjbb2013', 'name' : 'critical', 'score' : '<critical>'})
|
|
174 |
return Test("SPECjbb2013", ['-jar', 'specjbb2013.jar', '-m', 'composite'] +
|
|
175 |
_noneAsEmptyList(benchArgs), [success], [], [matcherCritical, matcherMax],
|
|
176 |
vmOpts=['-Xmx6g', '-Xms6g', '-Xmn3g', '-XX:+UseParallelOldGC', '-XX:-UseAdaptiveSizePolicy', '-XX:-UseBiasedLocking', '-XX:-UseCompressedOops'], defaultCwd=specjbb2013)
|
|
177 |
|
|
178 |
def getSPECjbb2015(benchArgs=None):
|
|
179 |
|
|
180 |
specjbb2015 = mx.get_env('SPECJBB2015')
|
|
181 |
if specjbb2015 is None or not exists(join(specjbb2015, 'specjbb2015.jar')):
|
|
182 |
mx.abort('Please set the SPECJBB2015 environment variable to a SPECjbb2015 directory')
|
|
183 |
|
|
184 |
jops = re.compile(r"^RUN RESULT: hbIR \(max attempted\) = [0-9]+, hbIR \(settled\) = [0-9]+, max-jOPS = (?P<max>[0-9]+), critical-jOPS = (?P<critical>[0-9]+)$", re.MULTILINE)
|
|
185 |
# error?
|
|
186 |
success = re.compile(r"org.spec.jbb.controller: Run finished", re.MULTILINE)
|
|
187 |
matcherMax = ValuesMatcher(jops, {'group' : 'SPECjbb2015', 'name' : 'max', 'score' : '<max>'})
|
|
188 |
matcherCritical = ValuesMatcher(jops, {'group' : 'SPECjbb2015', 'name' : 'critical', 'score' : '<critical>'})
|
|
189 |
return Test("SPECjbb2015", ['-jar', 'specjbb2015.jar', '-m', 'composite'] +
|
|
190 |
_noneAsEmptyList(benchArgs), [success], [], [matcherCritical, matcherMax],
|
|
191 |
vmOpts=['-Xmx6g', '-Xms6g', '-Xmn3g', '-XX:+UseParallelOldGC', '-XX:-UseAdaptiveSizePolicy', '-XX:-UseBiasedLocking', '-XX:-UseCompressedOops'], defaultCwd=specjbb2015)
|
|
192 |
|
|
193 |
def getSPECjvm2008(benchArgs=None):
|
|
194 |
|
|
195 |
specjvm2008 = mx.get_env('SPECJVM2008')
|
|
196 |
if specjvm2008 is None or not exists(join(specjvm2008, 'SPECjvm2008.jar')):
|
|
197 |
mx.abort('Please set the SPECJVM2008 environment variable to a SPECjvm2008 directory')
|
|
198 |
|
|
199 |
score = re.compile(r"^(Score on|Noncompliant) (?P<benchmark>[a-zA-Z0-9\._]+)( result)?: (?P<score>[0-9]+((,|\.)[0-9]+)?)( SPECjvm2008 Base)? ops/m$", re.MULTILINE)
|
|
200 |
error = re.compile(r"^Errors in benchmark: ", re.MULTILINE)
|
|
201 |
# The ' ops/m' at the end of the success string is important : it's how you can tell valid and invalid runs apart
|
|
202 |
success = re.compile(r"^(Noncompliant c|C)omposite result: [0-9]+((,|\.)[0-9]+)?( SPECjvm2008 (Base|Peak))? ops/m$", re.MULTILINE)
|
|
203 |
matcher = ValuesMatcher(score, {'group' : 'SPECjvm2008', 'name' : '<benchmark>', 'score' : '<score>'})
|
|
204 |
|
|
205 |
return Test("SPECjvm2008", ['-jar', 'SPECjvm2008.jar'] + _noneAsEmptyList(benchArgs), [success], [error], [matcher], vmOpts=['-Xms3g', '-XX:+' + gc, '-XX:-UseCompressedOops'], defaultCwd=specjvm2008)
|
|
206 |
|
|
207 |
def getDacapos(level=SanityCheckLevel.Normal, gateBuildLevel=None, dacapoArgs=None, extraVmArguments=None):
|
|
208 |
checks = []
|
|
209 |
|
|
210 |
for (bench, ns) in dacapoSanityWarmup.items():
|
|
211 |
if ns[level] > 0:
|
|
212 |
if gateBuildLevel is None or gateBuildLevel in dacapoGateBuildLevels[bench]:
|
|
213 |
checks.append(getDacapo(bench, ['-n', str(ns[level])] + _noneAsEmptyList(dacapoArgs), extraVmArguments=extraVmArguments))
|
|
214 |
|
|
215 |
return checks
|
|
216 |
|
|
217 |
def getDacapo(name, dacapoArgs=None, extraVmArguments=None):
|
|
218 |
dacapo = mx.get_env('DACAPO_CP')
|
|
219 |
if dacapo is None:
|
|
220 |
l = mx.library('DACAPO', False)
|
|
221 |
if l is not None:
|
|
222 |
dacapo = l.get_path(True)
|
|
223 |
else:
|
|
224 |
mx.abort('DaCapo 9.12 jar file must be specified with DACAPO_CP environment variable or as DACAPO library')
|
|
225 |
|
|
226 |
if not isfile(dacapo) or not dacapo.endswith('.jar'):
|
|
227 |
mx.abort('Specified DaCapo jar file does not exist or is not a jar file: ' + dacapo)
|
|
228 |
|
|
229 |
dacapoSuccess = re.compile(r"^===== DaCapo 9\.12 ([a-zA-Z0-9_]+) PASSED in ([0-9]+) msec =====", re.MULTILINE)
|
|
230 |
dacapoFail = re.compile(r"^===== DaCapo 9\.12 ([a-zA-Z0-9_]+) FAILED (warmup|) =====", re.MULTILINE)
|
|
231 |
dacapoTime = re.compile(r"===== DaCapo 9\.12 (?P<benchmark>[a-zA-Z0-9_]+) PASSED in (?P<time>[0-9]+) msec =====")
|
|
232 |
dacapoTime1 = re.compile(r"===== DaCapo 9\.12 (?P<benchmark>[a-zA-Z0-9_]+) completed warmup 1 in (?P<time>[0-9]+) msec =====")
|
|
233 |
|
|
234 |
dacapoMatcher = ValuesMatcher(dacapoTime, {'group' : 'DaCapo', 'name' : '<benchmark>', 'score' : '<time>'})
|
|
235 |
dacapoMatcher1 = ValuesMatcher(dacapoTime1, {'group' : 'DaCapo-1stRun', 'name' : '<benchmark>', 'score' : '<time>'})
|
|
236 |
|
|
237 |
# Use ipv4 stack for dacapos; tomcat+solaris+ipv6_interface fails (see also: JDK-8072384)
|
|
238 |
return Test("DaCapo-" + name, ['-jar', mx._cygpathU2W(dacapo), name] + _noneAsEmptyList(dacapoArgs), [dacapoSuccess], [dacapoFail],
|
|
239 |
[dacapoMatcher, dacapoMatcher1],
|
|
240 |
['-Xms2g', '-XX:+' + gc, '-XX:-UseCompressedOops', "-Djava.net.preferIPv4Stack=true", '-G:+ExitVMOnException'] +
|
|
241 |
_noneAsEmptyList(extraVmArguments))
|
|
242 |
|
|
243 |
def getScalaDacapos(level=SanityCheckLevel.Normal, gateBuildLevel=None, dacapoArgs=None, extraVmArguments=None):
|
|
244 |
checks = []
|
|
245 |
|
|
246 |
for (bench, ns) in dacapoScalaSanityWarmup.items():
|
|
247 |
if ns[level] > 0:
|
|
248 |
if gateBuildLevel is None or gateBuildLevel in dacapoScalaGateBuildLevels[bench]:
|
|
249 |
checks.append(getScalaDacapo(bench, ['-n', str(ns[level])] + _noneAsEmptyList(dacapoArgs), extraVmArguments=extraVmArguments))
|
|
250 |
|
|
251 |
return checks
|
|
252 |
|
|
253 |
def getScalaDacapo(name, dacapoArgs=None, extraVmArguments=None):
|
|
254 |
dacapo = mx.get_env('DACAPO_SCALA_CP')
|
|
255 |
if dacapo is None:
|
|
256 |
l = mx.library('DACAPO_SCALA', False)
|
|
257 |
if l is not None:
|
|
258 |
dacapo = l.get_path(True)
|
|
259 |
else:
|
|
260 |
mx.abort('Scala DaCapo 0.1.0 jar file must be specified with DACAPO_SCALA_CP environment variable or as DACAPO_SCALA library')
|
|
261 |
|
|
262 |
if not isfile(dacapo) or not dacapo.endswith('.jar'):
|
|
263 |
mx.abort('Specified Scala DaCapo jar file does not exist or is not a jar file: ' + dacapo)
|
|
264 |
|
|
265 |
dacapoSuccess = re.compile(r"^===== DaCapo 0\.1\.0(-SNAPSHOT)? ([a-zA-Z0-9_]+) PASSED in ([0-9]+) msec =====", re.MULTILINE)
|
|
266 |
dacapoFail = re.compile(r"^===== DaCapo 0\.1\.0(-SNAPSHOT)? ([a-zA-Z0-9_]+) FAILED (warmup|) =====", re.MULTILINE)
|
|
267 |
dacapoTime = re.compile(r"===== DaCapo 0\.1\.0(-SNAPSHOT)? (?P<benchmark>[a-zA-Z0-9_]+) PASSED in (?P<time>[0-9]+) msec =====")
|
|
268 |
|
|
269 |
dacapoMatcher = ValuesMatcher(dacapoTime, {'group' : "Scala-DaCapo", 'name' : '<benchmark>', 'score' : '<time>'})
|
|
270 |
|
|
271 |
return Test("Scala-DaCapo-" + name, ['-jar', mx._cygpathU2W(dacapo), name] + _noneAsEmptyList(dacapoArgs), [dacapoSuccess], [dacapoFail], [dacapoMatcher], ['-Xms2g', '-XX:+' + gc, '-XX:-UseCompressedOops'] + _noneAsEmptyList(extraVmArguments))
|
|
272 |
|
|
273 |
def getBootstraps():
|
|
274 |
time = re.compile(r"Bootstrapping Graal\.+ in (?P<time>[0-9]+) ms( \(compiled (?P<methods>[0-9]+) methods\))?")
|
|
275 |
scoreMatcher = ValuesMatcher(time, {'group' : 'Bootstrap', 'name' : 'BootstrapTime', 'score' : '<time>'})
|
|
276 |
methodMatcher = ValuesMatcher(time, {'group' : 'Bootstrap', 'name' : 'BootstrapMethods', 'score' : '<methods>'})
|
|
277 |
scoreMatcherBig = ValuesMatcher(time, {'group' : 'Bootstrap-bigHeap', 'name' : 'BootstrapTime', 'score' : '<time>'})
|
|
278 |
methodMatcherBig = ValuesMatcher(time, {'group' : 'Bootstrap-bigHeap', 'name' : 'BootstrapMethods', 'score' : '<methods>'})
|
|
279 |
|
|
280 |
tests = []
|
|
281 |
tests.append(Test("Bootstrap", ['-version'], successREs=[time], scoreMatchers=[scoreMatcher, methodMatcher], ignoredVMs=['client', 'server'], benchmarkCompilationRate=False))
|
|
282 |
tests.append(Test("Bootstrap-bigHeap", ['-version'], successREs=[time], scoreMatchers=[scoreMatcherBig, methodMatcherBig], vmOpts=['-Xms2g'], ignoredVMs=['client', 'server'], benchmarkCompilationRate=False))
|
|
283 |
return tests
|
|
284 |
|
|
285 |
class CTWMode:
|
|
286 |
Full, NoInline = range(2)
|
|
287 |
|
|
288 |
def getCTW(vm, mode):
|
|
289 |
time = re.compile(r"CompileTheWorld : Done \([0-9]+ classes, [0-9]+ methods, (?P<time>[0-9]+) ms\)")
|
|
290 |
scoreMatcher = ValuesMatcher(time, {'group' : 'CompileTheWorld', 'name' : 'CompileTime', 'score' : '<time>'})
|
|
291 |
|
|
292 |
jre = os.environ.get('JAVA_HOME')
|
|
293 |
if exists(join(jre, 'jre')):
|
|
294 |
jre = join(jre, 'jre')
|
|
295 |
rtjar = join(jre, 'lib', 'rt.jar')
|
|
296 |
|
|
297 |
|
|
298 |
args = ['-XX:+CompileTheWorld', '-Xbootclasspath/p:' + rtjar]
|
|
299 |
if vm == 'jvmci':
|
|
300 |
args += ['-XX:+BootstrapGraal']
|
|
301 |
if mode >= CTWMode.NoInline:
|
|
302 |
if not mx_graal.isJVMCIEnabled(vm):
|
|
303 |
args.append('-XX:-Inline')
|
|
304 |
else:
|
|
305 |
args.append('-G:CompileTheWordConfig=-Inline')
|
|
306 |
|
|
307 |
return Test("CompileTheWorld", args, successREs=[time], scoreMatchers=[scoreMatcher], benchmarkCompilationRate=False)
|
|
308 |
|
|
309 |
|
|
310 |
class Tee:
|
|
311 |
def __init__(self):
|
|
312 |
self.output = StringIO.StringIO()
|
|
313 |
def eat(self, line):
|
|
314 |
self.output.write(line)
|
|
315 |
sys.stdout.write(line)
|
|
316 |
|
|
317 |
"""
|
|
318 |
Encapsulates a single program that is a sanity test and/or a benchmark.
|
|
319 |
"""
|
|
320 |
class Test:
|
|
321 |
def __init__(self, name, cmd, successREs=None, failureREs=None, scoreMatchers=None, vmOpts=None, defaultCwd=None, ignoredVMs=None, benchmarkCompilationRate=False):
|
|
322 |
|
|
323 |
self.name = name
|
|
324 |
self.successREs = _noneAsEmptyList(successREs)
|
|
325 |
self.failureREs = _noneAsEmptyList(failureREs) + [re.compile(r"Exception occurred in scope: ")]
|
|
326 |
self.scoreMatchers = _noneAsEmptyList(scoreMatchers)
|
|
327 |
self.vmOpts = _noneAsEmptyList(vmOpts)
|
|
328 |
self.cmd = cmd
|
|
329 |
self.defaultCwd = defaultCwd
|
|
330 |
self.ignoredVMs = _noneAsEmptyList(ignoredVMs)
|
|
331 |
self.benchmarkCompilationRate = benchmarkCompilationRate
|
|
332 |
if benchmarkCompilationRate:
|
|
333 |
self.vmOpts = self.vmOpts + ['-XX:+CITime']
|
|
334 |
|
|
335 |
def __str__(self):
|
|
336 |
return self.name
|
|
337 |
|
|
338 |
def test(self, vm, cwd=None, extraVmOpts=None, vmbuild=None):
|
|
339 |
"""
|
|
340 |
Run this program as a sanity test.
|
|
341 |
"""
|
|
342 |
if vm in self.ignoredVMs:
|
|
343 |
return True
|
|
344 |
if cwd is None:
|
|
345 |
cwd = self.defaultCwd
|
|
346 |
parser = OutputParser()
|
|
347 |
jvmError = re.compile(r"(?P<jvmerror>([A-Z]:|/).*[/\\]hs_err_pid[0-9]+\.log)")
|
|
348 |
parser.addMatcher(ValuesMatcher(jvmError, {'jvmError' : '<jvmerror>'}))
|
|
349 |
|
|
350 |
for successRE in self.successREs:
|
|
351 |
parser.addMatcher(ValuesMatcher(successRE, {'passed' : '1'}))
|
|
352 |
for failureRE in self.failureREs:
|
|
353 |
parser.addMatcher(ValuesMatcher(failureRE, {'failed' : '1'}))
|
|
354 |
|
|
355 |
tee = Tee()
|
|
356 |
retcode = mx_graal.run_vm(self.vmOpts + _noneAsEmptyList(extraVmOpts) + self.cmd, vm, nonZeroIsFatal=False, out=tee.eat, err=subprocess.STDOUT, cwd=cwd, debugLevel=vmbuild)
|
|
357 |
output = tee.output.getvalue()
|
|
358 |
valueMaps = parser.parse(output)
|
|
359 |
|
|
360 |
if len(valueMaps) == 0:
|
|
361 |
return False
|
|
362 |
|
|
363 |
record = {}
|
|
364 |
for valueMap in valueMaps:
|
|
365 |
for key, value in valueMap.items():
|
|
366 |
if record.has_key(key) and record[key] != value:
|
|
367 |
mx.abort('Inconsistant values returned by test machers : ' + str(valueMaps))
|
|
368 |
record[key] = value
|
|
369 |
|
|
370 |
jvmErrorFile = record.get('jvmError')
|
|
371 |
if jvmErrorFile:
|
|
372 |
mx.log('/!\\JVM Error : dumping error log...')
|
|
373 |
with open(jvmErrorFile, 'rb') as fp:
|
|
374 |
mx.log(fp.read())
|
|
375 |
os.unlink(jvmErrorFile)
|
|
376 |
return False
|
|
377 |
|
|
378 |
if record.get('failed') == '1':
|
|
379 |
return False
|
|
380 |
|
|
381 |
return retcode == 0 and record.get('passed') == '1'
|
|
382 |
|
|
383 |
def bench(self, vm, cwd=None, extraVmOpts=None, vmbuild=None):
|
|
384 |
"""
|
|
385 |
Run this program as a benchmark.
|
|
386 |
"""
|
|
387 |
if vm in self.ignoredVMs:
|
|
388 |
return {}
|
|
389 |
if cwd is None:
|
|
390 |
cwd = self.defaultCwd
|
|
391 |
parser = OutputParser()
|
|
392 |
|
|
393 |
for successRE in self.successREs:
|
|
394 |
parser.addMatcher(ValuesMatcher(successRE, {'passed' : '1'}))
|
|
395 |
for failureRE in self.failureREs:
|
|
396 |
parser.addMatcher(ValuesMatcher(failureRE, {'failed' : '1'}))
|
|
397 |
for scoreMatcher in self.scoreMatchers:
|
|
398 |
parser.addMatcher(scoreMatcher)
|
|
399 |
|
|
400 |
if self.benchmarkCompilationRate:
|
|
401 |
if vm == 'jvmci':
|
|
402 |
bps = re.compile(r"ParsedBytecodesPerSecond@final: (?P<rate>[0-9]+)")
|
|
403 |
ibps = re.compile(r"InlinedBytecodesPerSecond@final: (?P<rate>[0-9]+)")
|
|
404 |
parser.addMatcher(ValuesMatcher(bps, {'group' : 'ParsedBytecodesPerSecond', 'name' : self.name, 'score' : '<rate>'}))
|
|
405 |
parser.addMatcher(ValuesMatcher(ibps, {'group' : 'InlinedBytecodesPerSecond', 'name' : self.name, 'score' : '<rate>'}))
|
|
406 |
else:
|
|
407 |
ibps = re.compile(r"(?P<compiler>[\w]+) compilation speed: +(?P<rate>[0-9]+) bytes/s {standard")
|
|
408 |
parser.addMatcher(ValuesMatcher(ibps, {'group' : 'InlinedBytecodesPerSecond', 'name' : '<compiler>:' + self.name, 'score' : '<rate>'}))
|
|
409 |
|
|
410 |
startDelim = 'START: ' + self.name
|
|
411 |
endDelim = 'END: ' + self.name
|
|
412 |
|
|
413 |
outputfile = os.environ.get('BENCH_OUTPUT', None)
|
|
414 |
if outputfile:
|
|
415 |
# Used only to debug output parsing
|
|
416 |
with open(outputfile) as fp:
|
|
417 |
output = fp.read()
|
|
418 |
start = output.find(startDelim)
|
|
419 |
end = output.find(endDelim, start)
|
|
420 |
if start == -1 and end == -1:
|
|
421 |
return {}
|
|
422 |
output = output[start + len(startDelim + os.linesep): end]
|
|
423 |
mx.log(startDelim)
|
|
424 |
mx.log(output)
|
|
425 |
mx.log(endDelim)
|
|
426 |
else:
|
|
427 |
tee = Tee()
|
|
428 |
mx.log(startDelim)
|
|
429 |
if mx_graal.run_vm(self.vmOpts + _noneAsEmptyList(extraVmOpts) + self.cmd, vm, nonZeroIsFatal=False, out=tee.eat, err=subprocess.STDOUT, cwd=cwd, debugLevel=vmbuild) != 0:
|
|
430 |
mx.abort("Benchmark failed (non-zero retcode)")
|
|
431 |
mx.log(endDelim)
|
|
432 |
output = tee.output.getvalue()
|
|
433 |
|
|
434 |
groups = {}
|
|
435 |
passed = False
|
|
436 |
for valueMap in parser.parse(output):
|
|
437 |
assert (valueMap.has_key('name') and valueMap.has_key('score') and valueMap.has_key('group')) or valueMap.has_key('passed') or valueMap.has_key('failed'), valueMap
|
|
438 |
if valueMap.get('failed') == '1':
|
|
439 |
mx.abort("Benchmark failed")
|
|
440 |
if valueMap.get('passed') == '1':
|
|
441 |
passed = True
|
|
442 |
groupName = valueMap.get('group')
|
|
443 |
if groupName:
|
|
444 |
group = groups.setdefault(groupName, {})
|
|
445 |
name = valueMap.get('name')
|
|
446 |
score = valueMap.get('score')
|
|
447 |
if name and score:
|
|
448 |
group[name] = score
|
|
449 |
|
|
450 |
if not passed:
|
|
451 |
mx.abort("Benchmark failed (not passed)")
|
|
452 |
|
|
453 |
return groups
|