This repository contains the case resources of the TTC 2017 Live contest.
The docs/2017_TTC_Live.pdf
file contains the case description.
- Python 2.7 or higher
- R
-
- jsonlite
-
- ggplot2
- .NET 4.5.1 to run correctness checks (the model printer uses NMF). .NET 4.5.1 in turn requires at least Windows Vista SP1.
- NMF: .NET Framework 4.5.1 and MSBuild on the path
The scripts
directory contains the run.py
script which is used for the following purposes:
run.py -b
-- builds the projectsrun.py -b -s
-- builds the projects without testingrun.py -m
-- runs the benchmarkrun.py -v
-- visualizes the results of the latest benchmark
The config
directory contains the configuration for the scripts:
config.json
-- configuration for the model generation and the benchmarkreporting.json
-- configuration for the visualization
The script runs the benchmark for the given number of runs, for the specified tools and models
The benchmark results are stored in a CSV file. The header for the CSV file is stored in the output/header.csv
file.
Make sure you read the README.md
file in the reporting
directory and install all the requirements for R.
To run the benchmark framework, simply run the run.py
script in the scripts folder with your Python interpreter. The working directory is not important.
To implement a tool, you need to
- create a new directory in the solutions directory and give it a suitable name
- implement the code generators :)
- add a
solution.ini
file that specifies the exact command that is run for your solution - add your tool to the configuration file
config.json
. The name you enter there must match the name of the directory created earlier.