## About the Tests

ARES-6 measures the execution time of JavaScript’s newest features, including symbols, for-of, arrow functions, Map/Set/WeakMap, let/const, classes, proxies, string interpolation, destructuring, default arguments, spread, tail calls, and generators. ARES-6 is comprised of four sub-tests: Air, Basic, Babylon, and ML.

Air is an ES2015 port of the WebKit B3 JIT’s `Air::allocateStack`

phase. This code is a heavy user of Map, Set, classes, spread, and for-of. The benchmark runs `allocateStack`

on hot function bodies from other popular JavaScript benchmarks: `executeIteration`

from Octane/Gameboy, `gaussianBlur`

from Kraken, and `scanIdentifier`

from Octane/Typescript. Air also runs `allocateStack`

on a hot function from Air itself. Browse the source.

Basic is an ES2015 implementation of the ECMA-55 BASIC standard. BASIC was an interactive language from the start, so our implementation uses generator functions for the entire interpreter. When the BASIC program says `INPUT`

, our interpreter says `yield`

. This allows users of the interpreter to easily integrate it into a runloop. Basic also uses classes, Map, and WeakMap. The benchmark runs a handful of simple programs, the most complex of which finds prime numbers. Browse the source.

Babylon is an implementation of a parser for the JavaScript language. Babylon is the parser used in the Babel JavaScript transpiler. The benchmark runs the Babylon parser on four different JavaScript sources. Babylon makes heavy use of classes, does non trivial string processing, and creates non-trivial object graphs.

ML is an implementation of a feedforward neural network. The benchmark trains several networks using different activation functions and several sample data sets. ML makes heavy use of classes. It relies on the ml-matrix library and does non-trivial matrix math.

ARES-6 rewards browsers that start up quickly and run smoothly. It's not enough to just measure the total running time of a workload. Browsers may perform differently for the same workload depending on how many times it has run. Garbage collection runs periodically, making some iterations take longer than others. Code that runs repeatedly gets optimized by the browser, so the first iteration of any workload is more expensive than the rest. ARES-6 runs Air, Basic, and Babylon for 200 iterations, and ML for 60 iterations. For each benchmark, ARES-6 reports the time of the first iteration, the average of the 4 worst iterations, and the average of all but the first iteration. ARES-6 also reports an overall score which is the geometric mean of all twelve of these numbers. ARES-6 equally rewards fast start-up, low jank, and sophisticated adaptive optimizations for long-running code.

Each ARES-6 sample has 200 iterations of Air, Basic, and Babylon, and 60 iterations of ML. ARES-6 runs 6 samples, and reports the average with 95% confidence intervals. Each sample runs in a fresh `iframe`

to simulate some of the effects of page navigation.