We demonstrate identification of position, material, orientation, and shape of objects imaged by a ^{85}Rb atomic magnetometer performing electromagnetic induction imaging supported by machine learning. Machine learning maximizes the information extracted from the images created by the magnetometer, demonstrating the use of hidden data. Localization 2.6 times better than the spatial resolution of the imaging system and successful classification up to 97% are obtained. This circumvents the need of solving the inverse problem and demonstrates the extension of machine learning to diffusive systems, such as low-frequency electrodynamics in media. Automated collection of task-relevant information from quantum-based electromagnetic imaging will have a relevant impact from biomedicine to security.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1103/PhysRevLett.120.033204 | DOI Listing |
Int J Med Inform
January 2025
School of Geography and the Environment, University of Oxford, South Parks Road, Oxford OX1 3QY, United Kingdom. Electronic address:
Background: Coronavirus Disease 2019 (COVID-19), caused by the SARS-CoV-2 virus, emerged as a global health crisis in 2019, resulting in widespread morbidity and mortality. A persistent challenge during the pandemic has been the accuracy of reported epidemic data, particularly in underdeveloped regions with limited access to COVID-19 test kits and healthcare infrastructure. In the post-COVID era, this issue remains crucial.
View Article and Find Full Text PDFInt J Med Inform
January 2025
Rheumatology and Allergy Clinical Epidemiology Research Center and Division of Rheumatology, Allergy, and Immunology, and Mongan Institute, Department of Medicine, Massachusetts General Hospital Boston MA USA. Electronic address:
Background: ANCA-associated vasculitis (AAV) is a rare but serious disease. Traditional case-identification methods using claims data can be time-intensive and may miss important subgroups. We hypothesized that a deep learning model analyzing electronic health records (EHR) can more accurately identify AAV cases.
View Article and Find Full Text PDFJMIR Cancer
January 2025
Division of Radiology and Biomedical Engineering, Graduate School of Medicine, The University of Tokyo, Tokyo, Japan.
Background: The application of natural language processing in medicine has increased significantly, including tasks such as information extraction and classification. Natural language processing plays a crucial role in structuring free-form radiology reports, facilitating the interpretation of textual content, and enhancing data utility through clustering techniques. Clustering allows for the identification of similar lesions and disease patterns across a broad dataset, making it useful for aggregating information and discovering new insights in medical imaging.
View Article and Find Full Text PDFBMC Pulm Med
January 2025
Universal Scientific Education and Research Network (USERN), Tehran, Iran.
Objective: Lung cancer (LC), the primary cause for cancer-related death globally is a diverse illness with various characteristics. Saliva is a readily available biofluid and a rich source of miRNA. It can be collected non-invasively as well as transported and stored easily.
View Article and Find Full Text PDFBMC Bioinformatics
January 2025
School of Computer Science and Technology, University of Science and Technology of China, 443 Huangshan Road, Hefei, 230027, China.
Background: Drug-drug interactions (DDIs) especially antagonistic ones present significant risks to patient safety, underscoring the urgent need for reliable prediction methods. Recently, substructure-based DDI prediction has garnered much attention due to the dominant influence of functional groups and substructures on drug properties. However, existing approaches face challenges regarding the insufficient interpretability of identified substructures and the isolation of chemical substructures.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!