TestEngineering/Performance/Talos/Misc

From MozillaWiki
< TestEngineering‎ | Performance‎ | Talos
Revision as of 15:41, 18 January 2018 by Igoldan (talk | contribs) (Created page with " = Adding a new test = Adding a new performance test or modifying an existing test is much easier than most people think. I general we need to create a patch for talos which...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Adding a new test

Adding a new performance test or modifying an existing test is much easier than most people think. I general we need to create a patch for talos which has:

  • determine if this is a startup test or a page load test and create the appropriate folder for your test in talos
  • add all files and resources for the test (make sure there is no external network accesses) to that folder
  • file a bug in the testing:talos component with your patch

What we need to know about tests

When adding a new test, we really need to understand what we are doing. Here are some questions that you should know the answer to before adding a new test:

  • What does this test measure?
  • Does this test overlap with any existing test?
  • What is the unit of measurement that we are recording?
  • What would constitute a regression?
  • What is the expected range in the results over time?
  • Are there variables or conditions which would affect this test?
    • browser configuration (prefs, environment variables)?
    • OS, resources, time of day, etc... ?
  • Indepenedent of Observation? Will this test produce the same number regardless of what was run before it?
  • What considerations are there for how this test should be run and what tools are required?

Please document the answers to these questions on Buildbot/Talos/Tests.

Making a Pageloader test work

A pageloader test is the most common. Using the pageloader extension allows for most of the reporting and talos integration to be done for you. Here is what you need to do:

  • Create a new directory in the tests subdirectory of talos
  • Create a manifest
    • add a <testname>.manifest file in your directory (svg example)
      • raw page load will just be a single url line (svg example)
      • internal measurements will have a % prior to the url (svg example)
  • Add your tests
  • Add your test definition to talos via [test.py] to add a new class for your test. ([tscrollx example])
    • Add an item for tpmanifest, ${talos} is replaced with the current running directory.
      • example: tpmanifest = '${talos}/tests/scroll/scroll.manifest'
    • Add an item for tpcycles (we recommend 1) and tppagecycles (we recommend 25 for a simple page, and 5 for a internal recording benchmark)
    • Add an item for filters. ([tcanvasmark example])

Making a Startup test work

A startup test is designed to load the browser with a single page many times. This is useful for tests which shouldn't have any extensions, can handle all loading and measuring themselves, or require the measurement of the browser startup. Here is what you need to do:

Steps to add a test to production

  • Update talos
    • just land code in-tree (inbound/fx-team)
  • Ensure the test follows the branch through the [release stages]
    • if we are dealing with buildbot configs, then we will need bugs for each merge cycle to turn on/off appropriate tests
    • if we are just updating talos.json, this will be automatic
  • if this is an update to an existing test that could change the numbers, this needs to be treated as a new test and run side by side for a week to get a new baseline for the numbers.
  • File a bug to get Treeherder updated with a new letter to track this test.
  • document the test on Buildbot/Talos and Buildbot/Talos/Tests
    • NOTE: this should be updated every time the branch moves closer to release. Once on release we don't need to update

While that is a laundry list of items to do, if you are developer of a component just talk to the a*team ([jmaher]) and they will handle the majority of the steps above.

Background Information

Hardware Profile of machines used in automation

  • linux32, linux64, winXP, win7, win8x64:
iX21X4 2U Neutron "Gemini" Series Four Node Hot-Pluggable Server (4 nodes per 2U)
920W High-Efficiency redundant (1+1) Power Supply
1 Intel X3450 CPU per node
8GB Total: 2 x 4GB DDR3 1333Mhz ECC/REG RAM per node
1 WD5003ABYX hard drive per node
1 NVIDIA GPU GeForce GT 610 per node
  • OSX 10.8 (mtnlion):
Model Name: Mac mini
Model Identifier: Macmini5,3
Processor Name: Intel Core i7
Processor Speed: 2 GHz
Number of Processors: 1
Total Number of Cores: 4
L2 Cache (per Core): 256 KB
L3 Cache: 6 MB
Memory: 8 GB
Disk: 2 x 500G SATA drives (RAID)
    • All Mac Minis have EDID devices attached that set the resolution at 1600x1200.

Naming convention

't' is pre-pended to the names to represent 'test'. Thus, ts = 'test startup', tp = 'test pageload', tdhtml = 'test dhtml'.

History of tp Tests

tp

The original tp test created by Mozilla to test browser page load time. Cycled through 40 pages. The pages were copied from the live web during November, 2000. Pages were cycled by loading them within the main browser window from a script that lived in content.

tp2/tp_js

The same tp test but loading the individual pages into a frame instead of the main browser window. Still used the old 40 page, year 2000 web page test set.

tp3

An update to both the page set and the method by which pages are cycled. The page set is now 393 pages from December, 2006. The pageloader is re-built as an extension that is pre-loaded into the browser chrome/components directories.

tp4

Updated web page test set to 100 pages from February 2009.

tp4m

This is a smaller pageset (21 pages) designed for mobile Firefox. This is a blend of regular and mobile friendly pages.

We landed on this on April 18th, 2011 in bug 648307. This runs for Android and Maemo mobile builds only.

tp5

Updated web page test set to 100 pages from April 8th, 2011. Effort was made for the pages to no longer be splash screens/login pages/home pages but to be pages that better reflect the actual content of the site in question.