Automatic summarization generation of sports video content has been object of great interest for many years. Although semantic descriptions techniques have been proposed, many of the approaches still rely on low-level video descriptors that render quite limited results due to the complexity of the problem and to the low capability of the descriptors to represent semantic content. In this paper, a new approach for automatic highlights summarization generation of soccer videos using audio-visual descriptors is presented. The approach is based on the segmentation of the video sequence into shots that will be further analyzed to determine its relevance and interest. Of special interest in the approach is the use of the audio information that provides additional robustness to the overall performance of the summarization system. For every video shot a set of low and mid level audio-visual descriptors are computed and lately adequately combined in order to obtain different relevance measures based on empirical knowledge rules. The final summary is generated by selecting those shots with highest interest according to the specifications of the user and the results of relevance measures. A variety of results are presented with real soccer video sequences that prove the validity of the approach.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4485672 | PMC |
http://dx.doi.org/10.1186/s40064-015-1065-9 | DOI Listing |
J Acoust Soc Am
April 2024
School of Architecture, Harbin Institute of Technology, Shenzhen, Shenzhen, 518000, China.
In the current study on soundscape, the distinction between felt emotion and perceived emotion in soundscape measurement has not been addressed as much as that in music studies. This research was conducted to investigate perceived and felt emotions associated with soundscape evaluation in urban open spaces through a laboratory audio-visual experiment using photographs and binaural recordings of 16 urban open locations across Harbin, China. In total, 46 participants were required to assess both the "perceived emotion" and "felt emotion" of the soundscapes using a questionnaire (in Chinese).
View Article and Find Full Text PDFJ Clin Med
January 2022
Penn Medicine Center for Connected Care, Hospital of the University of Pennsylvania, Philadelphia, PA 19104, USA.
A 24/7 telemedicine respiratory therapist (eRT) service was set up as part of the established University of Pennsylvania teleICU (PENN E-LERT) service during the COVID-19 pandemic, serving five hospitals and 320 critical care beds to deliver effective remote care in lieu of a unit-based RT. The eRT interventions were components of an evidence-based care bundle and included ventilator liberation protocols, low tidal volume protocols, tube patency, and an extubation checklist. In addition, the proactive rounding of patients, including ventilator checks, was included.
View Article and Find Full Text PDFMed Biol Eng Comput
February 2022
Department of Evolutionary Anthropology, University of Vienna, Althan strasse 14, A-1090, Vienna, Austria.
Investigating gender differences based on emotional changes using electroencephalogram (EEG) is essential to understand various human behavior in the individual situation in our daily life. However, gender differences based on EEG and emotional states are not thoroughly investigated. The main novelty of this paper is twofold.
View Article and Find Full Text PDFSpringerplus
July 2015
Signal Theory and Communications Department, UPC-BARCELONATECH, Esteve Terradas, 7, 08860 Castelldefels, Spain.
Automatic summarization generation of sports video content has been object of great interest for many years. Although semantic descriptions techniques have been proposed, many of the approaches still rely on low-level video descriptors that render quite limited results due to the complexity of the problem and to the low capability of the descriptors to represent semantic content. In this paper, a new approach for automatic highlights summarization generation of soccer videos using audio-visual descriptors is presented.
View Article and Find Full Text PDFJ Acoust Soc Am
November 2014
School of the Built Environment, Heriot-Watt University, Edinburgh EH14 4AS, United Kingdom.
This paper examines the audio-visual interaction and perception of water features used over road traffic noise, including their semantic aural properties, as well as their categorization and evocation properties. The research focused on a wide range of small to medium sized water features that can be used in gardens and parks to promote peacefulness and relaxation. Paired comparisons highlighted the inter-dependence between uni-modal (audio-only or visual-only) and bi-modal (audio-visual) perception, indicating that equal attention should be given to the design of both stimuli.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!