The web provides access to millions of datasets that can have additional impact when used beyond their original context. We have little empirical insight into what makes a dataset more reusable than others and which of the existing guidelines and frameworks, if any, make a difference. In this paper, we explore potential reuse features through a literature review and present a case study on datasets on GitHub, a popular open platform for sharing code and data. We describe a corpus of more than 1.4 million data files, from over 65,000 repositories. Using GitHub's engagement metrics as proxies for dataset reuse, we relate them to reuse features from the literature and devise an initial model, using deep neural networks, to predict a dataset's reusability. This demonstrates the practical gap between principles and actionable insights that allow data publishers and tools designers to implement functionalities that provably facilitate reuse.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7691392PMC
http://dx.doi.org/10.1016/j.patter.2020.100136DOI Listing

Publication Analysis

Top Keywords

dataset reuse
8
reuse features
8
features literature
8
reuse translating
4
translating principles
4
principles practice
4
practice web
4
web access
4
access millions
4
millions datasets
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!