Publications by authors named "P Radeva"

Background: Approximately 1 in 6 cannabis users develop a cannabis use disorder (CUD) and the odds increase to 1 in 2 for daily users.

Objective: The Dual use of Cannabis and Tobacco Monitoreing through a Gamified Web app (DuCATA_GAM-CaT) project aims to identify cannabis-tobacco patterns of use and withdrawal symptoms among individuals with CUD who are attending substance abuse programs.

Methods: The project uses a mixed methods approach consisting of 3 studies.

View Article and Find Full Text PDF
Article Synopsis
  • Explainable artificial intelligence (XAI) helps us understand how AI makes decisions, which is important for trusting its predictions.
  • A review of XAI used in heart-related AI shows only 37% of studies checked the quality of explanations, with many not evaluating them at all.
  • The goal is to encourage more research in healthcare to not just use XAI, but also assess its explanations to ensure the AI is safe and reliable.
View Article and Find Full Text PDF

Explainable Artificial Intelligence (XAI) provides tools to help understanding how AI models work and reach a particular decision or outcome. It helps to increase the interpretability of models and makes them more trustworthy and transparent. In this context, many XAI methods have been proposed to make black-box and complex models more digestible from a human perspective.

View Article and Find Full Text PDF

The outbreak of COVID-19 has shocked the entire world with its fairly rapid spread, and has challenged different sectors. One of the most effective ways to limit its spread is the early and accurate diagnosing of infected patients. Medical imaging, such as X-ray and computed tomography (CT), combined with the potential of artificial intelligence (AI), plays an essential role in supporting medical personnel in the diagnosis process.

View Article and Find Full Text PDF

Leveraging inexpensive and human intervention-based annotating methodologies, such as crowdsourcing and web crawling, often leads to datasets with noisy labels. Noisy labels can have a detrimental impact on the performance and generalization of deep neural networks. Robust models that are able to handle and mitigate the effect of these noisy labels are thus essential.

View Article and Find Full Text PDF