PeerJ Computer Science (Apr 2017)

Interoperability and FAIRness through a novel combination of Web technologies

  • Mark D. Wilkinson,
  • Ruben Verborgh,
  • Luiz Olavo Bonino da Silva Santos,
  • Tim Clark,
  • Morris A. Swertz,
  • Fleur D.L. Kelpin,
  • Alasdair J.G. Gray,
  • Erik A. Schultes,
  • Erik M. van Mulligen,
  • Paolo Ciccarese,
  • Arnold Kuzniar,
  • Anand Gavai,
  • Mark Thompson,
  • Rajaram Kaliyaperumal,
  • Jerven T. Bolleman,
  • Michel Dumontier

DOI
https://doi.org/10.7717/peerj-cs.110
Journal volume & issue
Vol. 3
p. e110

Abstract

Read online Read online

Data in the life sciences are extremely diverse and are stored in a broad spectrum of repositories ranging from those designed for particular data types (such as KEGG for pathway data or UniProt for protein data) to those that are general-purpose (such as FigShare, Zenodo, Dataverse or EUDAT). These data have widely different levels of sensitivity and security considerations. For example, clinical observations about genetic mutations in patients are highly sensitive, while observations of species diversity are generally not. The lack of uniformity in data models from one repository to another, and in the richness and availability of metadata descriptions, makes integration and analysis of these data a manual, time-consuming task with no scalability. Here we explore a set of resource-oriented Web design patterns for data discovery, accessibility, transformation, and integration that can be implemented by any general- or special-purpose repository as a means to assist users in finding and reusing their data holdings. We show that by using off-the-shelf technologies, interoperability can be achieved atthe level of an individual spreadsheet cell. We note that the behaviours of this architecture compare favourably to the desiderata defined by the FAIR Data Principles, and can therefore represent an exemplar implementation of those principles. The proposed interoperability design patterns may be used to improve discovery and integration of both new and legacy data, maximizing the utility of all scholarly outputs.

Keywords