The field of natural language processing (NLP) has seen rapid advances in the past several years since the introduction of deep learning techniques. A variety of NLP tasks including syntactic parsing, machine translation, and summarization can now be performed by relatively simple combinations of general neural network models such as recurrent neural networks and attention mechanisms. This manuscript gives a brief introduction to deep learning and an overview of the current deep learning-based NLP technology.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.11477/mf.1416201215 | DOI Listing |
Radiology
January 2025
From the Department of Diagnostic and Interventional Radiology, University Hospital Bonn, Venusberg-Campus 1, 53127 Bonn, Germany.
Background Large-scale secondary use of clinical databases requires automated tools for retrospective extraction of structured content from free-text radiology reports. Purpose To share data and insights on the application of privacy-preserving open-weights large language models (LLMs) for reporting content extraction with comparison to standard rule-based systems and the closed-weights LLMs from OpenAI. Materials and Methods In this retrospective exploratory study conducted between May 2024 and September 2024, zero-shot prompting of 17 open-weights LLMs was preformed.
View Article and Find Full Text PDFVis Intell
December 2024
Department of Information Technology and Electrical Engineering, ETH Zurich, Sternwartstrasse 7, Zürich, Switzerland.
The LLaMA family, a collection of foundation language models ranging from 7B to 65B parameters, has become one of the most powerful open-source large language models (LLMs) and the popular LLM backbone of multi-modal large language models (MLLMs), widely used in computer vision and natural language understanding tasks. In particular, LLaMA3 models have recently been released and have achieved impressive performance in various domains with super-large scale pre-training on over 15T tokens of data. Given the wide application of low-bit quantization for LLMs in resource-constrained scenarios, we explore LLaMA3's capabilities when quantized to low bit-width.
View Article and Find Full Text PDFBMJ Open
December 2024
Postgraduate Program in Public Health, Federal University of Rio Grande do Norte, Natal, Brazil.
Introduction: The global phenomenon of population ageing is expanding rapidly, with WHO projecting that the number of individuals aged 60 and over will reach 2.1 billion by 2050, a significant rise from 900 million in 2015. This pronounced growth poses substantial challenges to healthcare systems globally, necessitating the development of effective public policies to ensure adequate access to healthcare services for the elderly demographic.
View Article and Find Full Text PDFSci Rep
January 2025
EIAS Data Science Lab, College of Computer and Information Sciences, Prince Sultan University, 11586, Riyadh, Saudi Arabia.
During the Covid-19 pandemic, the widespread use of social media platforms has facilitated the dissemination of information, fake news, and propaganda, serving as a vital source of self-reported symptoms related to Covid-19. Existing graph-based models, such as Graph Neural Networks (GNNs), have achieved notable success in Natural Language Processing (NLP). However, utilizing GNN-based models for propaganda detection remains challenging because of the challenges related to mining distinct word interactions and storing nonconsecutive and broad contextual data.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!