IEEE/ACM Trans Comput Biol Bioinform
October 2024
Machine learning algorithms have been extensively used for accurate classification of cancer subtypes driven by gene expression-based biomarkers. However, biomarker models combining multiple gene expression signatures are often not reproducible in external validation datasets and their feature set size is often not optimized, jeopardizing their translatability into cost-effective clinical tools. We investigated how to solve the multi-objective problem of finding the best trade-offs between classification performance and set size applying seven algorithms for machine learning-driven feature subset selection and analyse how they perform in a benchmark with eight large-scale transcriptome datasets of cancer, covering both training and external validation sets.
View Article and Find Full Text PDFBioinform Adv
October 2022
Motivation: Gene expression-based classifiers are often developed using historical data by training a model on a small set of patients and a large set of features. Models trained in such a way can be afterwards applied for predicting the output for new unseen patient data. However, very often the accuracy of these models starts to decrease as soon as new data is fed into the trained model.
View Article and Find Full Text PDFMotivation: In modern translational research, the development of biomarkers heavily relies on use of omics technologies, but implementations with basic data mining algorithms frequently lead to false positives. Non-dominated Sorting Genetic Algorithm II (NSGA2) is an extremely effective algorithm for biomarker discovery but has been rarely evaluated against large-scale datasets. The exploration of the feature search space is the key to NSGA2 success but in specific cases NSGA2 expresses a shallow exploration of the space of possible feature combinations, possibly leading to models with low predictive performances.
View Article and Find Full Text PDFAm J Geriatr Psychiatry
September 2022
Objective: To develop streamlined Risk Prediction Models (Manto RPMs) for late-life depression.
Design: Prospective study.
Setting: The Survey of Health, Ageing and Retirement in Europe (SHARE) study.
Comput Struct Biotechnol J
March 2022
The recent advancements in toxicogenomics have led to the availability of large omics data sets, representing the starting point for studying the exposure mechanism of action and identifying candidate biomarkers for toxicity prediction. The current lack of standard methods in data generation and analysis hampers the full exploitation of toxicogenomics-based evidence in regulatory risk assessment. Moreover, the pipelines for the preprocessing and downstream analyses of toxicogenomic data sets can be quite challenging to implement.
View Article and Find Full Text PDFThe pharmacological arsenal against the COVID-19 pandemic is largely based on generic anti-inflammatory strategies or poorly scalable solutions. Moreover, as the ongoing vaccination campaign is rolling slower than wished, affordable and effective therapeutics are needed. To this end, there is increasing attention toward computational methods for drug repositioning and de novo drug design.
View Article and Find Full Text PDFDNA microarrays are widely used to investigate gene expression. Even though the classical analysis of microarray data is based on the study of differentially expressed genes, it is well known that genes do not act individually. Network analysis can be applied to study association patterns of the genes in a biological system.
View Article and Find Full Text PDFThe amount of data made available by microarrays gives researchers the opportunity to delve into the complexity of biological systems. However, the noisy and extremely high-dimensional nature of this kind of data poses significant challenges. Microarrays allow for the parallel measurement of thousands of molecular objects spanning different layers of interactions.
View Article and Find Full Text PDFBiomarkers are valuable indicators of the state of a biological system. Microarray technology has been extensively used to identify biomarkers and build computational predictive models for disease prognosis, drug sensitivity and toxicity evaluations. Activation biomarkers can be used to understand the underlying signaling cascades, mechanisms of action and biological cross talk.
View Article and Find Full Text PDFPreprocessing of transcriptomics data plays a pivotal role in the development of toxicogenomics-driven tools for chemical toxicity assessment. The generation and exploitation of large volumes of molecular profiles, following an appropriate experimental design, allows the employment of toxicogenomics (TGx) approaches for a thorough characterisation of the mechanism of action (MOA) of different compounds. To date, a plethora of data preprocessing methodologies have been suggested.
View Article and Find Full Text PDFThe starting point of successful hazard assessment is the generation of unbiased and trustworthy data. Conventional toxicity testing deals with extensive observations of phenotypic endpoints in vivo and complementing in vitro models. The increasing development of novel materials and chemical compounds dictates the need for a better understanding of the molecular changes occurring in exposed biological systems.
View Article and Find Full Text PDFTranscriptomics data are relevant to address a number of challenges in Toxicogenomics (TGx). After careful planning of exposure conditions and data preprocessing, the TGx data can be used in predictive toxicology, where more advanced modelling techniques are applied. The large volume of molecular profiles produced by omics-based technologies allows the development and application of artificial intelligence (AI) methods in TGx.
View Article and Find Full Text PDFBackground: Identifying those people at increased risk of early functional decline in activities of daily living (ADL) is essential for initiating preventive interventions. The aim of this study is to develop and validate a clinical prediction model for onset of functional decline in ADL in three years of follow-up in older people of 65-75 years old.
Methods: Four population-based cohort studies were pooled for the analysis: ActiFE-ULM (Germany), ELSA (United Kingdom), InCHIANTI (Italy), LASA (Netherlands).
Assessing the risk to develop a specific disease is the first step towards prevention, both at individual and population levels. The development and validation of risk prediction models (RPMs) is the norm within different fields of medicine but still underused in psychiatry, despite the global impact of mental disorders. In particular, there is a lack of RPMs to assess the risk of developing depression, the first worldwide cause of disability and harbinger of functional decline in old age.
View Article and Find Full Text PDFBackground: Early identification of people at risk of functional decline is essential for delivering targeted preventive interventions.
Objective: The aim of this study is to identify and predict trajectories of functional decline over 9 years in males and females aged 60-70 years.
Methods: We included 403 community-dwelling participants from the InCHIANTI study and 395 from the LASA study aged 60-70 years at baseline, of whom the majority reported no functional decline at baseline (median 0, interquartile range 0-1).
Background And Objective: The fall risk assessment tool (FRAT-up) is a tool for predicting falls in community-dwelling older people based on a meta-analysis of fall risk factors. Based on the fall risk factor profile, this tool calculates the individual risk of falling over the next year. The objective of this study is to evaluate the performance of FRAT-up in predicting future falls in multiple cohorts.
View Article and Find Full Text PDFBackground: About 30% of people over 65 are subject to at least one unintentional fall a year. Fall prevention protocols and interventions can decrease the number of falls. To be effective, a prevention strategy requires a prior step to evaluate the fall risk of the subjects.
View Article and Find Full Text PDF