Skip to content

newlawrence/Parsers-Benchmark

 
 

Repository files navigation

The Great C++ Mathematical Expression Parser Benchmark

This is a benchmark suite for different implementations of open source math expression parsers and evaluators written in C++. Currently the following expression parsers are part of this benchmark:

# Library Author License Numeric Type
00 ATMSP Heinz van Saanen GPL v3 double, MPFR
01 ExprTk Arash Partow MIT double, float, MPFR
02 FParser Juha Nieminen & Joel Yliluoma LGPL double
03 Lepton Peter Eastman MIT double
04 MathExpr Yann Ollivier Copyright Notice 1997-2000 double
05 MTParser Mathieu Jacques CPOL double
06 muParser Ingo Berg MIT double, float
07 muParserX Ingo Berg MIT double, float
08 TinyExpr Lewis Van Winkle Zlib double
09 Calculate Alberto Lorenzo Márquez MIT double {complex}

Note: The terms double, float etc found in the table above are defined as follows:

Note: This archive is not meant as a primary source for any of the libraries listed above as it may contain outdated versions of said libraries.

Objectives

All the parsers in this benchmark suite are in one way or another POEM (Parse Once Evaluate Many times) in nature. As such the primary objectives for this benchmark are to ascertain the following properties for each of the denoted parsers in following order of priority:

  1. Correctness and precision of results relative to the floating point type used
  2. Robustness of the parser given various mathematical expressions
  3. Evaluation efficiency of 'compiled/parsed' expressions

The Setup

The benchmark takes two parameters:

  1. Number of iterations (N)
  2. Benchmark Test File

The mathematical expression benchmark files are as follows:

  1. bench_expr.txt
  2. bench_expr_all.txt
  3. bench_expr_all_permutations.txt
  4. bench_expr_complete.txt
  5. bench_expr_extensive.txt
  6. bench_expr_precedence.txt
  7. bench_expr_random_with_functions.txt
  8. bench_expr_random_without_functions.txt
  9. bench_expr_weird.txt

The following is a short sampling of expressions taken from the bench_expr.txt expression suite:

  • a+1.1
  • a*2.2
  • 2.2*a+1.1
  • (2.2*a+1.1)*3.3
  • sin(2.2 * a) + cos(pi / b)
  • 1 - sin(2.2 * a) + cos(pi / b)
  • sqrt(1 - sin(2.2 * a) + cos(pi / b) / 3.3)
  • (a^2 / sin(2 * pi / b)) -a / 2.2
  • 1-(a/b*0.5)
  • e^log(7*a)
  • 10^log(3+b)
  • (cos(2.41)/b)
  • -(sin(pi+a)+1)
  • a-(e^(log(7+b)))

As an example, the following execution will use the"bench_expr_all.txt" set of expressions, executing each expression 100000 times:

ParserBench.exe 100000 bench_expr_all.txt

The Rounds

For every expression in the benchmark file, every parser evaluates the given expression N times, this is known as a round. The total time each parser takes to evaluate the expression N times is recorded. Ranking of the parsers for the round is done from the fastest to the slowest. The following is an example summary for the expression '(a^2/sin(2*pi/b))-a/2':

Expression 64 of 70: "(a^2/sin(2*pi/b))-a/2"; Progress: ############
[01] ExprTk              ( 35.863 ns, 3.744853294789905362, -15768714.840563148260116577)
[02] ExprTkFloat         ( 37.413 ns, 3.744852304458618164, -15768714.994192123413085938)
[03] muparserSSE         ( 44.672 ns, 3.744852304458618164, -15768715.739250183105468750)
[04] muparser 2.2.4 (omp)( 50.093 ns, 3.744853294789905362, -15768714.840563148260116577)
[05] MTParser            ( 54.681 ns, 3.744853294789905362, -15768714.840563148260116577)
[06] muparser 2.2.4      ( 55.816 ns, 3.744853294789905362, -15768714.840563148260116577)
[07] FParser 4.5         ( 65.141 ns, 3.744853294789905362, -15768714.840563148260116577)
[08] atmsp 1.0.4         ( 78.809 ns, 3.744853294789905362, -15768714.840563148260116577)
[09] MathExpr            (131.242 ns, 3.744853294789905362, -15768714.840563148260116577)
[10] TinyExpr            (168.341 ns, 3.744853294789905362, -15768714.840563148260116577)
[11] Lepton              (402.011 ns, 3.744853294789905362, -15768714.840563148260116577)
[12] muparserx           (572.611 ns, 3.744853294789905362, -15768714.840563148260116577)

Parsers that can't parse the given expression or produce a result different to the expected result (based on a normalized epsilon approach) are disqualified for the round and do not participate in point accumulation. The following is an example where a few of the parsers failed to either parse or evaluate the given expression of '+1.1+a^b' correctly:

Expression 63 of 96: "+a^+b+1.1"; Progress: ############
[01] ExprTk              ( 40.631 ns, 2.333286300554663129, 14534421.395495440810918808)
[02] ExprTkFloat         ( 44.212 ns, 2.333286285400390625, 14534422.159194946289062500)
[03] muparserSSE         ( 44.556 ns, 2.333286285400390625, 14534422.159194946289062500)
[04] atmsp 1.0.4         ( 47.906 ns, 2.333286300554663129, 14534421.395495440810918808)
[05] TinyExpr            ( 52.831 ns, 2.333286300554663129, 14534421.395495440810918808)
[06] MTParser            ( 56.045 ns, 2.333286300554663129, 14534421.395495440810918808)
[07] muparser 2.2.4      ( 67.061 ns, 2.333286300554663129, 14534421.395495440810918808)
[08] muparser 2.2.4 (omp)( 67.231 ns, 2.333286300554663129, 14534421.395495440810918808)
[09] muparserx           (251.338 ns, 2.333286300554663129, 14534421.395495440810918808)
DNQ List
[01] FParser 4.5         (Syntax error)
[02] MathExpr            (parsing error)
[03] Lepton              (Parse error in expression "+a^+b+1.1 ": unexpected token: +)
**** ERROR ****   Excessive number of evaluation failures!  [3]

The meanings of each of the columns are as follows:

  1. Rank
  2. Parser
  3. Average time (in nano-seconds) to evaluate the expression once, taken over N-times
  4. Result of evaluating the expression
  5. Sum of N-evaluations of the given expression

The Final Summary

Once all the expressions have been completed, a summary is provided, that includes information relating to the build parameters of the benchmark binary, the architecture which the benchmark was run upon, a final ranking of the parsers, and a listing of expressions per parser that were deemed as having been disqualified during the benchmark.

The following is an example summary generated at the completion of the benchmark:

Benchmark settings:
  - Expressions File is "bench_expr.txt"
  - Reference parser is ExprTk
  - Iterations per expression: 5000000
  - Number of expressions: 70
  - Number of excessive failures: 5
  - Release build
  - Compiled with MSVC Version 1900
  - IEEE 754 (IEC 559) is Available
  - 32-bit build
  - CPU Name      : Intel(R) Xeon(R) CPU E5-2687W 0 @ 3.10GHz
  - Num. cores    : 16
  - Num. logical  : 32
  - Total logical : 16
  - L1 D cache    : 32 KB
  - L1 I cache    : 32 KB
  - L2 cache      : 256 KB
  - L3 cache      : 20480 KB
  - L1D assoc.    : 8-way
  - L2 assoc.     : 8-way
  - L3 assoc.     : 20-way
  - L1D line size : 64 bytes
  - L2 line size  : 64 bytes
  - L3 line size  : 64 bytes
  - CPU clock     : 3092 MHz
  - Features      : acpi aes apic avx clflush cmov cx16 cx8 dca de ds_cpl
                    dts dts64 est fpu fxsr ht lahf_lm lm mca mce mmx monitor
                    msr mtrr osxsave pae pat pbe pclmul pdcm pge pni popcnt
                    pse pse36 sep smx ss sse sse2 sse4_1 sse4_2 ssse3 tm tm2
                    tsc vme vmx xd xsave xtpr

Scores:
#  Parser               Type       Points   Score    Failures
-------------------------------------------------------------
00 ExprTk               double     781      100      0
01 ExprTkFloat          float      643       82      9
02 muparserSSE          float      639       93      9
03 FParser 4.5          double     536       45      0
04 atmsp 1.0.4          double     482       40      2
05 muparser 2.2.4       double     445       37      0
06 MTParser             double     402       38     10
07 muparser 2.2.4 (omp) double     374       35      0
08 MathExpr             double     345       30      2
09 TinyExpr             double     296       30      2
10 Lepton               double     111        6      2
11 muparserx            double      97        5      0

The Results

The summaries of runs of the benchmark suite upon various architectures can be found here: Results

Example Results

The following is a chart depicting the results as executions per second obtained by running the 'bench_expr_random_without_functions.txt' benchmark, using the double type, 5000000 iterations per expression on an Intel Xeon E5-2687W 3GHz. The horizontal measure on the chart is the length of the expression in bytes.

ScreenShot

Note: The parsers muParserSSE and ExprTKFloat were excluded due to their use of the float type, and the Lepton and muParserX parsers were excluded due to the fact that their results were consistently 2-3 orders of magnitude slower than all the others.

Note: The command executed to generate the above results is as follows:

ParserBenchmark.exe 5000000 bench_expr_random_without_functions.txt write_table

Native Mode

The benchmark has a special mode called "Native". In this mode a predefined set of expressions are employed, and benchmarked against both native equivalent implementations and the standard parsers/evaluators.

The command to execute the benchmark in native mode is as follows:

ParserBenchmark.exe 1000000 native

Customised Tests

Creating a customised test is very simple. All that is required is to create a utf-8 encoded text file, and pass it as the 2nd parameter to the benchmark executable.

Each line can only have one expression, comprised of the following:

  • Operators: +, -, *, /, ^
  • Variables: a, b, c, x, y, z, w
  • Constants: e and pi
  • Functions: sin, cos, tan, abs, exp, sqrt, log, pow
  • Parentheses: Round '(' and ')'

Furthermore blank lines or lines beginning with a #-symbol will be ignored (aka comments or white-space).

The following is a simple example of a custom test:

---- snip ----

# My first expression:
(x + 1) / 2 * b

# Pythagoras
a^2 + b^2

# A constant expression
1 + (2 * 3) / 7

# Definition of tan in terms of sine and cosine
sin(x) / cos(x)

---- snip ----

New Parsers

The addition of new or unlisted C++ mathematical expression parsers is always welcome. If you know of a parser and would like it included in the suite feel free to make a request. Additionally if the associated benchmark implementation (eg: src/BenchXYZW.cpp) could be made available as part of the request that would very helpful.

About

C++ Mathematical Expression Parser Benchmark

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C++ 86.5%
  • C 10.7%
  • HTML 1.0%
  • Shell 0.5%
  • Makefile 0.4%
  • C# 0.3%
  • Other 0.6%