138 lines
		
	
	
		
			5.4 KiB
		
	
	
	
		
			Python
		
	
	
	
			
		
		
	
	
			138 lines
		
	
	
		
			5.4 KiB
		
	
	
	
		
			Python
		
	
	
	
| # Copyright (c) 2012 The Chromium OS Authors. All rights reserved.
 | |
| # Use of this source code is governed by a BSD-style license that can be
 | |
| # found in the LICENSE file.
 | |
| #
 | |
| # GLMark outputs a final performance score, and it checks the performance score
 | |
| # against minimum requirement if min_score is set.
 | |
| 
 | |
| import logging
 | |
| import os
 | |
| import re
 | |
| import string
 | |
| 
 | |
| from autotest_lib.client.bin import test, utils
 | |
| from autotest_lib.client.common_lib import error
 | |
| from autotest_lib.client.cros import service_stopper
 | |
| from autotest_lib.client.cros.graphics import graphics_utils
 | |
| 
 | |
| GLMARK2_TEST_RE = (
 | |
|     r'^\[(?P<scene>.*)\] (?P<options>.*): FPS: (?P<fps>\d+) FrameTime: '
 | |
|     r'(?P<frametime>\d+.\d+) ms$')
 | |
| GLMARK2_SCORE_RE = r'glmark2 Score: (\d+)'
 | |
| 
 | |
| # perf value description strings may only contain letters, numbers, periods,
 | |
| # dashes and underscores.
 | |
| # But glmark2 test names are usually in the form:
 | |
| #   scene-name:opt=val:opt=v1,v2;v3,v4 or scene:<default>
 | |
| # which we convert to:
 | |
| #   scene-name.opt_val.opt_v1-v2_v3-v4 or scene.default
 | |
| description_table = string.maketrans(':,=;', '.-__')
 | |
| description_delete = '<>'
 | |
| 
 | |
| 
 | |
| class graphics_GLMark2(graphics_utils.GraphicsTest):
 | |
|     """Runs glmark2, which benchmarks only calls compatible with OpenGL ES 2.0"""
 | |
|     version = 1
 | |
|     preserve_srcdir = True
 | |
|     _services = None
 | |
| 
 | |
|     def setup(self):
 | |
|         self.job.setup_dep(['glmark2'])
 | |
| 
 | |
|     def initialize(self):
 | |
|         super(graphics_GLMark2, self).initialize()
 | |
|         # If UI is running, we must stop it and restore later.
 | |
|         self._services = service_stopper.ServiceStopper(['ui'])
 | |
|         self._services.stop_services()
 | |
| 
 | |
|     def cleanup(self):
 | |
|         if self._services:
 | |
|             self._services.restore_services()
 | |
|         super(graphics_GLMark2, self).cleanup()
 | |
| 
 | |
|     @graphics_utils.GraphicsTest.failure_report_decorator('graphics_GLMark2')
 | |
|     def run_once(self, size='800x600', hasty=False, min_score=None):
 | |
|         dep = 'glmark2'
 | |
|         dep_dir = os.path.join(self.autodir, 'deps', dep)
 | |
|         self.job.install_pkg(dep, 'dep', dep_dir)
 | |
| 
 | |
|         glmark2 = os.path.join(self.autodir, 'deps/glmark2/glmark2')
 | |
|         if not os.path.exists(glmark2):
 | |
|             raise error.TestFail('Failed: Could not find test binary.')
 | |
| 
 | |
|         glmark2_data = os.path.join(self.autodir, 'deps/glmark2/data')
 | |
| 
 | |
|         options = []
 | |
|         options.append('--data-path %s' % glmark2_data)
 | |
|         options.append('--size %s' % size)
 | |
|         options.append('--annotate')
 | |
|         if hasty:
 | |
|             options.append('-b :duration=0.2')
 | |
|         else:
 | |
|             options.append('-b :duration=2')
 | |
|         cmd = glmark2 + ' ' + ' '.join(options)
 | |
| 
 | |
|         if os.environ.get('CROS_FACTORY'):
 | |
|             from autotest_lib.client.cros import factory_setup_modules
 | |
|             from cros.factory.test import ui
 | |
|             ui.start_reposition_thread('^glmark')
 | |
| 
 | |
|         # TODO(ihf): Switch this test to use perf.PerfControl like
 | |
|         #            graphics_GLBench once it is stable. crbug.com/344766.
 | |
|         if not hasty:
 | |
|             if not utils.wait_for_idle_cpu(60.0, 0.1):
 | |
|                 if not utils.wait_for_idle_cpu(20.0, 0.2):
 | |
|                     raise error.TestFail('Failed: Could not get idle CPU.')
 | |
|             if not utils.wait_for_cool_machine():
 | |
|                 raise error.TestFail('Failed: Could not get cold machine.')
 | |
| 
 | |
|         # In this test we are manually handling stderr, so expected=True.
 | |
|         # Strangely autotest takes CmdError/CmdTimeoutError as warning only.
 | |
|         try:
 | |
|             result = utils.run(cmd,
 | |
|                                stderr_is_expected=True,
 | |
|                                stdout_tee=utils.TEE_TO_LOGS,
 | |
|                                stderr_tee=utils.TEE_TO_LOGS)
 | |
|         except error.CmdError:
 | |
|             raise error.TestFail('Failed: CmdError running %s' % cmd)
 | |
|         except error.CmdTimeoutError:
 | |
|             raise error.TestFail('Failed: CmdTimeout running %s' % cmd)
 | |
| 
 | |
|         logging.info(result)
 | |
|         for line in result.stderr.splitlines():
 | |
|             if line.startswith('Error:'):
 | |
|                 # Line already starts with 'Error: ", not need to prepend.
 | |
|                 raise error.TestFail(line)
 | |
| 
 | |
|         # Numbers in hasty mode are not as reliable, so don't send them to
 | |
|         # the dashboard etc.
 | |
|         if not hasty:
 | |
|             keyvals = {}
 | |
|             score = None
 | |
|             # glmark2 output the final performance score as:
 | |
|             #  glmark2 Score: 530
 | |
|             for line in result.stdout.splitlines():
 | |
|                 match = re.findall(GLMARK2_SCORE_RE, line)
 | |
|                 if match:
 | |
|                     score = int(match[0])
 | |
|             if not score:
 | |
|                 raise error.TestFail('Failed: Unable to read benchmark score')
 | |
|             # Output numbers for plotting by harness.
 | |
|             logging.info('GLMark2 score: %d', score)
 | |
|             if os.environ.get('CROS_FACTORY'):
 | |
|                 from autotest_lib.client.cros import factory_setup_modules
 | |
|                 from cros.factory.event_log import EventLog
 | |
|                 EventLog('graphics_GLMark2').Log('glmark2_score', score=score)
 | |
|             keyvals['glmark2_score'] = score
 | |
|             self.write_perf_keyval(keyvals)
 | |
|             self.output_perf_value(
 | |
|                 description='Score',
 | |
|                 value=score,
 | |
|                 units='score',
 | |
|                 higher_is_better=True)
 | |
| 
 | |
|             if min_score is not None and score < min_score:
 | |
|                 raise error.TestFail(
 | |
|                     'Failed: Benchmark score %d < %d (minimum score '
 | |
|                     'requirement)' % (score, min_score))
 |