Medical image captioning provides the visual information of medical images in the form of natural language. It requires an efficient approach to understand and evaluate the similarity between visual and textual elements and to generate a sequence of output words. A novel show, attend, and tell model (ATM) is implemented, which considers a visual attention approach using an encoder-decoder model. But the show, attend, and tell model is sensitive to its initial parameters. Therefore, a Strength Pareto Evolutionary Algorithm-II (SPEA-II) is utilized to optimize the initial parameters of the ATM. Finally, experiments are considered using the benchmark data sets and competitive medical image captioning techniques. Performance analysis shows that the SPEA-II-based ATM performs significantly better as compared to the existing models.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8947912PMC
http://dx.doi.org/10.1155/2022/9638438DOI Listing

Publication Analysis

Top Keywords

medical image
12
image captioning
12
attend model
8
initial parameters
8
medical
4
captioning optimized
4
optimized deep
4
deep learning
4
model
4
learning model
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!