Visual dialog demonstrates several important aspects of multimodal artificial intelligence; however, it is hindered by visual grounding and visual coreference resolution problems. To overcome these problems, we propose the novel neural module network for visual dialog (NMN-VD). NMN-VD is an efficient question-customized modular network model that combines only the modules required for deciding answers after analyzing input questions. In particular, the model includes a module that effectively finds the visual area indicated by a pronoun using a reference pool to solve a visual coreference resolution problem, which is an important challenge in visual dialog. In addition, the proposed NMN-VD model includes a method for distinguishing and handling impersonal pronouns that do not require visual coreference resolution from general pronouns. Furthermore, a new module that effectively handles comparison questions found in visual dialogs is included in the model, as well as a module that applies a triple-attention mechanism to solve visual grounding problems between the question and the image. The results of various experiments conducted using a set of large-scale benchmark data verify the efficacy and high performance of our proposed NMN-VD model.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7866498 | PMC |
http://dx.doi.org/10.3390/s21030931 | DOI Listing |
J Clin Psychol
January 2025
Department of Psychiatry and Psychotherapy, University Medical Center Hamburg-Eppendorf, Hamburg, Germany.
Treatment efficacy for patients with obsessive-compulsive disorder (OCD) with poor insight is low. Insight refers to a patient's ability to recognize that their obsessions are irrational and that their compulsions are futile attempts to reduce anxiety. This case study presents the first application of virtual reality-assisted avatar therapy for OCD (VRT-OCD) in a patient with contamination OCD and ambivalent insight.
View Article and Find Full Text PDFJ Clin Psychol
January 2025
Department of Clinical Psychology and Psychobiology, The Institute of Neurosciences, Universitat de Barcelona, Barcelona, Spain.
Based on the repertory grid technique, we developed Explore Your Meanings (EYME), a digital platform that helps patients explore identity values and internal conflicts using virtual reality (VR). EYME was part of a research project treating depression in young adults, including 10 weekly, 1-h sessions aimed at changing personal constructs-cognitive schemas that shape how individuals interpret reality. We present the case of Mary, a 21-year-old woman diagnosed with persistent major depressive disorder and social phobia.
View Article and Find Full Text PDFFront Psychol
October 2024
Department of Communication Sciences and Disorders, University of Mississippi, Oxford, MS, United States.
The purpose of the current study was to expand upon previous research on RECALL, a dialogic reading intervention modified for autistic children aimed at increasing engagement. Children ages 3-6 years ( = 6) with language delays with or without co-occurring autism were tested using a multiple baseline across participants design. During baseline, the interventionist used dialogic reading and asked questions after every page.
View Article and Find Full Text PDFNucleic Acids Res
January 2025
Genomics Institute, University of California Santa Cruz, Santa Cruz, CA 95064, USA.
The UCSC Genome Browser (https://genome.ucsc.edu) is a widely utilized web-based tool for visualization and analysis of genomic data, encompassing over 4000 assemblies from diverse organisms.
View Article and Find Full Text PDFIn clinical guidelines for patients with chronic musculoskeletal pain, reassurance is a key element. The purpose of reassuring patients is to change their views on their illness and, thereby, their actions. However, when symptoms persist without pathological findings, reassurance can be difficult to achieve.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!