There is a numerical benchmark reported in Fig. 4 with absolute runtimes and memory usages that can directly be reproduced with the provided source code. The benchmark was performed on the author's computer, and since numerical performance and parallel scaling can be somewhat hardware-dependent, it would be of interest to see whether a performance that is comparable to the one reported in the paper can be reproduced by others on their own computers in practice. The benchmark simulates a growing tissue from one to 10,000 cells in just ten minutes, so this offers an easy entry point into tissue modeling and simulation. No input data is needed to reproduce the output. The program has no dependencies.
I tried hard to make it reproducible, so hopefully this paper can serve as an example on how reproducibility can be achieved. I think that being reproducible with only few commands to type in a terminal is quite an achievment. At least in my field, where I usually see code published along with paper, but with almost no documentation on how to rerun it.
Because: - Two fellow PhDs working on different topics have been able to reproduce some figures by following the README instructions and I hope this extends to other people - I've tried to incorporate as many of the best practices as possible to make my code and data open and accessible - I've tried to make sure that my data is exactly reproducible with the specified random seed strategy - the paper suggests a method that should be useful to other researchers in my field, which is not useful unless my results are reproducible
The original data took quite a while to produce for a previous paper, but for this paper, all tables and figures should be exactly reproducible by simply running the jupyter notebook.