The emergence of nanoinformatics as a key component of nanotechnology and nanosafety assessment for the prediction of engineered nanomaterials (NMs) properties, interactions, and hazards, and for grouping and read-across to reduce reliance on animal testing, has put the spotlight firmly on the need for access to high-quality, curated datasets. To date, the focus has been around what constitutes data quality and completeness, on the development of minimum reporting standards, and on the FAIR (findable, accessible, interoperable, and reusable) data principles. However, moving from the theoretical realm to practical implementation requires human intervention, which will be facilitated by the definition of clear roles and responsibilities across the complete data lifecycle and a deeper appreciation of what metadata is, and how to capture and index it. Here, we demonstrate, using specific worked case studies, how to organise the nano-community efforts to define metadata schemas, by organising the data management cycle as a joint effort of all players (data creators, analysts, curators, managers, and customers) supervised by the newly defined role of data shepherd. We propose that once researchers understand their tasks and responsibilities, they will naturally apply the available tools. Two case studies are presented (modelling of particle agglomeration for dose metrics, and consensus for NM dissolution), along with a survey of the currently implemented metadata schema in existing nanosafety databases. We conclude by offering recommendations on the steps forward and the needed workflows for metadata capture to ensure FAIR nanosafety data.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7602672 | PMC |
http://dx.doi.org/10.3390/nano10102033 | DOI Listing |
Gigascience
January 2025
INRAE, Université de Bordeaux, F-33610 Cestas, France.
Background: Descriptive metadata are vital for reporting, discovering, leveraging, and mobilizing research datasets. However, resolving metadata issues as part of a data management plan can be complex for data producers. To organize and document data, various descriptive metadata must be created.
View Article and Find Full Text PDFPharmacoepidemiol Drug Saf
December 2024
Department of Data Science and Biostatistics, Julius Center for Health Science and Primary Care, University Medical Center of Utrecht, Utrecht University, Utrecht, The Netherlands.
Objective: To enhance documentation on programming decisions in Real World Evidence (RWE) studies.
Materials And Methods: We analyzed several statistical analysis plans (SAP) within the Vaccine Monitoring Collaboration for Europe (VAC4EU) to identify study design sections and specifications for programming RWE studies. We designed a machine-readable metadata schema containing study sections, codelists, and time anchoring definitions specified in the SAPs with adaptability and user-friendliness.
Patterns (N Y)
November 2024
Forschungszentrum Jülich GmbH, Institute of Climate and Energy Systems (ICE) - Jülich Systems Analysis (ICE-2), 52425 Jülich, Germany.
The reuse of research software is central to research efficiency and academic exchange. The application of software enables researchers to reproduce, validate, and expand upon study findings. The analysis of open-source code aids in the comprehension, comparison, and integration of approaches.
View Article and Find Full Text PDFBMC Med Inform Decis Mak
November 2024
Medical Informatics Group, Center of Health Data Science, Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Charitéplatz 1, 10117, Berlin, Germany.
Background: Clinical data warehouses provide harmonized access to healthcare data for medical researchers. Informatics for Integrating Biology and the Bedside (i2b2) is a well-established open-source solution with the major benefit that data representations can be tailored to support specific use cases. These data representations can be defined and improved via an iterative approach together with domain experts and the medical researchers using the platform.
View Article and Find Full Text PDFOpen Res Eur
October 2024
Computing and Automatics Department, Campus Viriato, Escuela Politécnica Superior de Zamora, Avenida de Requejo, 33, Universidad de Salamanca, Zamora, 49022, Spain.
Background: Earth Observation (EO) datasets have become vital for decision support applications, particularly from open satellite portals that provide extensive historical datasets. These datasets can be integrated with in-situ data to power artificial intelligence mechanisms for accurate forecasting and trend analysis. However, researchers and data scientists face challenges in finding appropriate EO datasets due to inconsistent metadata structures and varied keyword descriptions.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!