Audio classification and retrieval has been recognized as a fascinating field of endeavor for as long as it has existed due to the topic of identifying and choosing the most useful audio attributes. The categorization of audio files is significant not only in the area of multimedia applications but also in the disciplines of medicine, sound analysis, intelligent homes and cities, urban informatics, entertainment, and surveillance. This study introduces a new algorithm called the modified bacterial foraging optimization algorithm (MBFOA), which is based on a method that retrieves and classifies audio data. The goal of this algorithm is to reduce the computational complexity of existing techniques. Along with the combination of the peak estimated signal, the enhanced mel-frequency cepstral coefficient (EMFCC) and the enhanced power normalized cepstral coefficients (EPNCC) are used. These are then optimized using the fitness function utilizing MBFOA. The probabilistic neural network is used to differentiate between a music signal and a spoken signal from an audio source (PNN). It is next necessary to extract and list the characteristics that correspond to the class that was arrived at as a consequence of the categorization. When compared to other approaches that are somewhat similar, MBFOA demonstrates superior levels of sensitivity, specificity, and accuracy.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC10348865 | PMC |
http://dx.doi.org/10.1155/2023/7735846 | DOI Listing |
Nat Comput Sci
January 2025
Key Lab of Fabrication Technologies for Integrated Circuits and Key Laboratory of Microelectronic Devices and Integrated Technology, Institute of Microelectronics of the Chinese Academy of Sciences, Beijing, China.
The human brain is a complex spiking neural network (SNN) capable of learning multimodal signals in a zero-shot manner by generalizing existing knowledge. Remarkably, it maintains minimal power consumption through event-based signal propagation. However, replicating the human brain in neuromorphic hardware presents both hardware and software challenges.
View Article and Find Full Text PDFSensors (Basel)
December 2024
Faculty of Medicine and Health Technology, Tampere University, 33720 Tampere, Finland.
Extracting behavioral information from animal sounds has long been a focus of research in bioacoustics, as sound-derived data are crucial for understanding animal behavior and environmental interactions. Traditional methods, which involve manual review of extensive recordings, pose significant challenges. This study proposes an automated system for detecting and classifying animal vocalizations, enhancing efficiency in behavior analysis.
View Article and Find Full Text PDFAAPS PharmSciTech
January 2025
OSIS, Silver Spring, Maryland, U.S.A.
Travel restrictions during the novel coronavirus, SARS-CoV-2 (COVID-19) public health emergency affected the U.S. Food and Drug Administration's (FDA) ability to conduct on-site bioavailability/bioequivalence (BA/BE) and Good Laboratory Practice (GLP) nonclinical inspections.
View Article and Find Full Text PDFConf Proc (IEEE Colomb Conf Commun Comput)
August 2024
School of Electronic and Electrical Engineering, Sungkyunkwan University, South Korea.
A new pneumonia detection method is proposed to provide both pneumonia detection in respiratory sound signals and wheeze and crackle discrimination when pneumonia episodes are detected. In the proposed method, two-step hierarchy, classifying pneumonia in the first step and discriminating wheezing and crackling in the second step, is considered; the conventional pneumonia detection method is modified to improve pneumonia detection performance, while wheezing and crackling discrimination functionality is added to facilitate the application of appropriate remedies for each case. We used resampling techniques to address the imbalance in the ICBHI pneumonia dataset.
View Article and Find Full Text PDFFront Artif Intell
December 2024
Computer Science and Software Engineering Department, Auckland University of Technology, Auckland, New Zealand.
Introduction: Musical instrument recognition is a critical component of music information retrieval (MIR), aimed at identifying and classifying instruments from audio recordings. This task poses significant challenges due to the complexity and variability of musical signals.
Methods: In this study, we employed convolutional neural networks (CNNs) to analyze the contributions of various spectrogram representations-STFT, Log-Mel, MFCC, Chroma, Spectral Contrast, and Tonnetz-to the classification of ten different musical instruments.
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!