Publications by authors named "Stephane Robin"

In a context of saturation of private dental practices and medical demography issues, responses to requests for emergency dental care are a poorly documented problem. In partnership with the Observatoire Regional de la Santé, the URPS Chirurgiens-Dentistes Nouvelle-Aquitaine, a union, conducted a survey of private dentists in May and June 2022. The objective was to estimate the volume of requests for unscheduled dental care and to describe the responses provided by professionals.

View Article and Find Full Text PDF

Next-generation biomonitoring proposes to combine machine-learning algorithms with environmental DNA data to automate the monitoring of the Earth's major ecosystems. In the present study, we searched for molecular biomarkers of tree water status to develop next-generation biomonitoring of forest ecosystems. Because phyllosphere microbial communities respond to both tree physiology and climate change, we investigated whether environmental DNA data from tree phyllosphere could be used as molecular biomarkers of tree water status in forest ecosystems.

View Article and Find Full Text PDF

The epidemic of Covid-19 was characterized, from its beginning, by "emergency". A state of emergency enacted by the state authorities to fight, on one hand, against the pandemic as such and, on the other hand, to manage the influx of patients admitted in intensive care. In this unprecedented context, the suffering of the people goes beyond the emergency situation and persists in forms ranging from a pseudo-banality to the complexity of an insidious evolution.

View Article and Find Full Text PDF

A lot of what we know about past speciation and extinction dynamics is based on statistically fitting birth-death processes to phylogenies of extant species. Despite their wide use, the reliability of these tools is regularly questioned. It was recently demonstrated that vast 'congruent' sets of alternative diversification histories cannot be distinguished (i.

View Article and Find Full Text PDF

Motivation: Combining the results of different experiments to exhibit complex patterns or to improve statistical power is a typical aim of data integration. The starting point of the statistical analysis often comes as a set of P-values resulting from previous analyses, that need to be combined flexibly to explore complex hypotheses, while guaranteeing a low proportion of false discoveries.

Results: We introduce the generic concept of composed hypothesis, which corresponds to an arbitrary complex combination of simple hypotheses.

View Article and Find Full Text PDF

The relative importance of ecological factors and species interactions for shaping species distributions is still debated. The realised niches of eight sympatric tephritid fruit flies were inferred from field abundance data using joint species distribution modelling and network inference, on the whole community and separately on three host plant groups. These estimates were then confronted the fundamental niches of seven fly species estimated through laboratory-measured fitnesses on host plants.

View Article and Find Full Text PDF

Motivation: De novo comparative metagenomics is one of the most straightforward ways to analyze large sets of metagenomic data. Latest methods use the fraction of shared k-mers to estimate genomic similarity between read sets. However, those methods, while extremely efficient, are still limited by computational needs for practical usage outside of large computing facilities.

View Article and Find Full Text PDF

Hidden Markov models provide a natural statistical framework for the detection of the copy number variations (CNV) in genomics. In this context, we define a hidden Markov process that underlies all individuals jointly in order to detect and to classify genomics regions in different states (typically, deletion, normal or amplification). Structural variations from different individuals may be dependent.

View Article and Find Full Text PDF

Omic data are characterized by the presence of strong dependence structures that result either from data acquisition or from some underlying biological processes. Applying statistical procedures that do not adjust the variable selection step to the dependence pattern may result in a loss of power and the selection of spurious variables. The goal of this paper is to propose a variable selection procedure within the multivariate linear model framework that accounts for the dependence between the multiple responses.

View Article and Find Full Text PDF

To study the evolution of several quantitative traits, the classical phylogenetic comparative framework consists of a multivariate random process running along the branches of a phylogenetic tree. The Ornstein-Uhlenbeck (OU) process is sometimes preferred to the simple Brownian motion (BM) as it models stabilizing selection toward an optimum. The optimum for each trait is likely to be changing over the long periods of time spanned by large modern phylogenies.

View Article and Find Full Text PDF

Background: Detecting local correlations in expression between neighboring genes along the genome has proved to be an effective strategy to identify possible causes of transcriptional deregulation in cancer. It has been successfully used to illustrate the role of mechanisms such as copy number variation (CNV) or epigenetic alterations as factors that may significantly alter expression in large chromosomal regions (gene silencing or gene activation).

Results: The identification of correlated regions requires segmenting the gene expression correlation matrix into regions of homogeneously correlated genes and assessing whether the observed local correlation is significantly higher than the background chromosomal correlation.

View Article and Find Full Text PDF

Background: Auxin is a major phytohormone involved in many developmental processes by controlling gene expression through a network of transcriptional regulators. In Arabidopsis thaliana, the auxin signalling network is made of 52 potentially interacting transcriptional regulators, activating or repressing gene expression. All the possible interactions were tested in two-way yeast-2-hybrid experiments.

View Article and Find Full Text PDF

Next-generation sequencing technologies now constitute a method of choice to measure gene expression. Data to analyze are read counts, commonly modeled using negative binomial distributions. A relevant issue associated with this probabilistic framework is the reliable estimation of the overdispersion parameter, reinforced by the limited number of replicates generally observable for each gene.

View Article and Find Full Text PDF

Dynamic extinction colonisation models (also called contact processes) are widely studied in epidemiology and in metapopulation theory. Contacts are usually assumed to be possible only through a network of connected patches. This network accounts for a spatial landscape or a social organization of interactions.

View Article and Find Full Text PDF

Background: Change point problems arise in many genomic analyses such as the detection of copy number variations or the detection of transcribed regions. The expanding Next Generation Sequencing technologies now allow to locate change points at the nucleotide resolution.

Results: Because of its complexity which is almost linear in the sequence length when the maximal number of segments is constant, and as its performance had been acknowledged for microarrays, we propose to use the Pruned Dynamic Programming algorithm for Seq-experiment outputs.

View Article and Find Full Text PDF

Background: The accurate prognosis definition to tailor treatment for early luminal invasive breast carcinoma patients remains challenging.

Materials And Methods: Two hundred fourteen early luminal breast carcinomas were genotyped with single nucleotide polymorphisms (SNPs) array to determine the number of chromosomal breakpoints as a marker of genomic instability. Proliferation was assessed by KI67 (immunohistochemistry) and genomic grade index (transcriptomic analysis).

View Article and Find Full Text PDF

The aim of this paper is to propose a test procedure for the detection of differential alternative splicing across conditions for tiling array or exon chip data. While developed in a mixed model framework, the test procedure is exact (avoiding computational burden) and applicable to a large variety of contrasts, including several previously published ones. A simulation study is presented to evaluate the robustness and performance of the method.

View Article and Find Full Text PDF

Post-translational modification of histones and DNA methylation are important components of chromatin-level control of genome activity in eukaryotes. However, principles governing the combinatorial association of chromatin marks along the genome remain poorly understood. Here, we have generated epigenomic maps for eight histone modifications (H3K4me2 and 3, H3K27me1 and 2, H3K36me3, H3K56ac, H4K20me1 and H2Bub) in the model plant Arabidopsis and we have combined these maps with others, produced under identical conditions, for H3K9me2, H3K9me3, H3K27me3 and DNA methylation.

View Article and Find Full Text PDF

The statistical analysis of array comparative genomic hybridization (CGH) data has now shifted to the joint assessment of copy number variations at the cohort level. Considering multiple profiles gives the opportunity to correct for systematic biases observed on single profiles, such as probe GC content or the so-called "wave effect." In this article, we extend the segmentation model developed in the univariate case to the joint analysis of multiple CGH profiles.

View Article and Find Full Text PDF

Tiling arrays make possible a large-scale exploration of the genome thanks to probes which cover the whole genome with very high density, up to 2,000,000 probes. Biological questions usually addressed are either the expression difference between two conditions or the detection of transcribed regions. In this work, we propose to consider both questions simultaneously as an unsupervised classification problem by modeling the joint distribution of the two conditions.

View Article and Find Full Text PDF
Tailored aggregation for classification.

IEEE Trans Pattern Anal Mach Intell

November 2009

Compression and variable selection are two classical strategies to deal with large-dimension data sets in classification. We propose an alternative strategy, called aggregation, which consists of a clustering step of redundant variables and a compression step within each group. We develop a statistical framework to define tailored aggregation methods that can be combined with selection methods to build reliable classifiers that benefit from the information contained in redundant variables.

View Article and Find Full Text PDF

Motivation: High-density oligonucleotide tiling array technology holds the promise of a better description of the complexity and the dynamics of transcriptional landscapes. In organisms such as bacteria and yeasts, transcription can be measured on a genome-wide scale with a resolution >25 bp. The statistical models currently used to handle these data remain however very simple, the most popular being the piecewise constant Gaussian model with a fixed number of breakpoints.

View Article and Find Full Text PDF

Background: As biological networks often show complex topological features, mathematical methods are required to extract meaningful information. Clustering methods are useful in this setting, as they allow the summary of the network's topology into a small number of relevant classes. Different strategies are possible for clustering, and in this article we focus on a model-based strategy that aims at clustering nodes based on their connectivity profiles.

View Article and Find Full Text PDF