The emergence of nanoinformatics as a key component of nanotechnology and nanosafety assessment for the prediction of engineered nanomaterials (NMs) properties, interactions, and hazards, and for grouping and read-across to reduce reliance on animal testing, has put the spotlight firmly on the need for access to high-quality, curated datasets. To date, the focus has been around what constitutes data quality and completeness, on the development of minimum reporting standards, and on the FAIR (findable, accessible, interoperable, and reusable) data principles. However, moving from the theoretical realm to practical implementation requires human intervention, which will be facilitated by the definition of clear roles and responsibilities across the complete data lifecycle and a deeper appreciation of what metadata is, and how to capture and index it. Here, we demonstrate, using specific worked case studies, how to organise the nano-community efforts to define metadata schemas, by organising the data management cycle as a joint effort of all players (data creators, analysts, curators, managers, and customers) supervised by the newly defined role of data shepherd. We propose that once researchers understand their tasks and responsibilities, they will naturally apply the available tools. Two case studies are presented (modelling of particle agglomeration for dose metrics, and consensus for NM dissolution), along with a survey of the currently implemented metadata schema in existing nanosafety databases. We conclude by offering recommendations on the steps forward and the needed workflows for metadata capture to ensure FAIR nanosafety data.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7602672PMC
http://dx.doi.org/10.3390/nano10102033DOI Listing

Publication Analysis

Top Keywords

metadata schemas
8
data
8
metadata capture
8
case studies
8
metadata
6
metadata stewardship
4
nanosafety
4
stewardship nanosafety
4
nanosafety community-driven
4
community-driven organisation
4

Similar Publications

Background: Descriptive metadata are vital for reporting, discovering, leveraging, and mobilizing research datasets. However, resolving metadata issues as part of a data management plan can be complex for data producers. To organize and document data, various descriptive metadata must be created.

View Article and Find Full Text PDF

Real-World Evidence BRIDGE: A Tool to Connect Protocol With Code Programming.

Pharmacoepidemiol Drug Saf

December 2024

Department of Data Science and Biostatistics, Julius Center for Health Science and Primary Care, University Medical Center of Utrecht, Utrecht University, Utrecht, The Netherlands.

Objective: To enhance documentation on programming decisions in Real World Evidence (RWE) studies.

Materials And Methods: We analyzed several statistical analysis plans (SAP) within the Vaccine Monitoring Collaboration for Europe (VAC4EU) to identify study design sections and specifications for programming RWE studies. We designed a machine-readable metadata schema containing study sections, codelists, and time anchoring definitions specified in the SAPs with adaptability and user-friendliness.

View Article and Find Full Text PDF

The reuse of research software is central to research efficiency and academic exchange. The application of software enables researchers to reproduce, validate, and expand upon study findings. The analysis of open-source code aids in the comprehension, comparison, and integration of approaches.

View Article and Find Full Text PDF

Background: Clinical data warehouses provide harmonized access to healthcare data for medical researchers. Informatics for Integrating Biology and the Bedside (i2b2) is a well-established open-source solution with the major benefit that data representations can be tailored to support specific use cases. These data representations can be defined and improved via an iterative approach together with domain experts and the medical researchers using the platform.

View Article and Find Full Text PDF

Enriching Earth observation datasets through semantics for climate change applications: The EIFFEL ontology.

Open Res Eur

October 2024

Computing and Automatics Department, Campus Viriato, Escuela Politécnica Superior de Zamora, Avenida de Requejo, 33, Universidad de Salamanca, Zamora, 49022, Spain.

Background: Earth Observation (EO) datasets have become vital for decision support applications, particularly from open satellite portals that provide extensive historical datasets. These datasets can be integrated with in-situ data to power artificial intelligence mechanisms for accurate forecasting and trend analysis. However, researchers and data scientists face challenges in finding appropriate EO datasets due to inconsistent metadata structures and varied keyword descriptions.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!