Papers



Submit a Paper!

Browse ReproHack papers

  • Optimizing the Use of Carbonate Standards to Minimize Uncertainties in Clumped Isotope Data

    Authors: Ilja J. Kocken, Inigo A. Müller, Martin Ziegler
    DOI: 10.1029/2019GC008545
    Submitted by japhir      

    Why should we attempt to reproduce this paper?

    Even though the approach in the paper focuses on a specific measurement (clumped isotopes) and how to optimize which and how many standards we use, I hope that the problem is general enough that insight can translate to any kind of measurement that relies on machine calibration. I've committed to writing a literate program (plain text interspersed with code chunks) to explain what is going on and to make the simulations one step at a time. I really hope that this is understandable to future collaborators and scientists in my field, but I have not had any code review internally and I also didn't receive any feedback on it from the reviewers. I would love to see if what in my mind represents "reproducible code" is actually reproducible, and to learn what I can improve for future projects!

  • Plasmonic nanostructure design and characterization via Deep Learning

    Authors: Malkiel, I., Mrejen, M., Nagler, A. et al.
    DOI: 10.1038/s41377-018-0060-7
    Submitted by hub-admin    

    Why should we attempt to reproduce this paper?

    The current code is written in Torch, which is no longer actively maintained. Since deep learning in nanophotonics is an area of active interest (e.g. for the design of new metamaterials), it is important to update the code to use a more modern deep learning library such as tensorflow/keras

  • Where should new parkrun events be located? Modelling the potential impact of 200 new events on socio-economic inequalities in access and participation

    Authors: Schneider PP, Smith RA, Bullas AM, Bayley T, Haake SS, Brennan A, Goyder E
    Submitted by hub-admin    
      Mean reproducibility score:   7.0/10   |   Number of reviews:   3
    Why should we attempt to reproduce this paper?

    If all went right, the analysis should be fully reproducible without the need to make any adjustments. The paper aims to find optimal locations for new parkruns, but we were not 100% sure how 'optimal' should be defined. We provide a few examples, but the code was meant to be flexible enough to allow potential decision makers to specify their own, alternative objectives. The spatial data set is also quite interesting and fun to play around with. Cave: The full analysis takes a while to run (~30+ min) and might require >= 8gb ram.

  • Open Trade Statistics

    Authors: Pachá (Mauricio Vargas Sepúlveda)
    Submitted by hub-admin    

    Why should we attempt to reproduce this paper?

    The focus of the project is reproducibility. Here we show the differences to access data compared to similar initiatives: https://ropensci.org/blog/2019/05/09/tradestatistics/. Also, similar projects have obscure parts, while our exposes the code from raw data downloading to dashboard creation.

    Tags: R Shiny