Retrieving nearest neighbors across correlated data in multiple modalities, such as image-text pairs on Facebook and video-tag pairs on YouTube, has become a challenging task due to the huge amount of data. Multimodal hashing methods that embed data into binary codes can boost the retrieving speed and reduce storage requirements. Since unsupervised multimodal hashing methods are usually inferior to supervised ones and the supervised ones require too much manually labeled data, the proposed method in this paper utilizes partially available labels to design a semisupervised multimodal hashing method. The labels for unlabeled data are treated as an interval type-2 fuzzy set and estimated by the available labels. By defuzzifying the estimated labels using hard partitioning, a supervised multimodal hashing method is used to generate binary codes. Experiments show that the proposed semisupervised method with 50% labels can get a medium performance among the compared supervised ones and achieve approximate performance to the best supervised method with 90% labels. With only 10% labels, the proposed method can still compete with the worst compared supervised one. Furthermore, the proposed label estimation method has been experimentally proven to be more feasible for a multilabeled MIRFlickr data set in a hash lookup task.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TCYB.2019.2906658 | DOI Listing |
Entropy (Basel)
October 2024
College of Computer Science and Technology, Xinjiang University, Urumqi 830046, China.
Deep hashing technology, known for its low-cost storage and rapid retrieval, has become a focal point in cross-modal retrieval research as multimodal data continue to grow. However, existing supervised methods often overlook noisy labels and multiscale features in different modal datasets, leading to higher information entropy in the generated hash codes and features, which reduces retrieval performance. The variation in text annotation information across datasets further increases the information entropy during text feature extraction, resulting in suboptimal outcomes.
View Article and Find Full Text PDFNeural Netw
February 2025
Big Data Institute, School of Computer Science and Engineering, Central South University, ChangSha, Hunan, 410000, China. Electronic address:
With the emergence of massive amounts of multi-source heterogeneous data on the Internet, quickly retrieving effective information from this extensive data has become a hot research topic. Due to the efficiency and speed of hash learning methods in multimedia retrieval, they have become a mainstream method for multimedia retrieval. However, unsupervised multimedia hash learning methods still face challenges with the difficulties of tuning due to the excessive number of hyperparameters and the lack of precise guidance on semantic similarity.
View Article and Find Full Text PDFBioengineering (Basel)
July 2024
National Genomics Data Center, China National Center for Bioinformation, Beijing 100101, China.
Unlabelled: In medical image retrieval, accurately retrieving relevant images significantly impacts clinical decision making and diagnostics. Traditional image-retrieval systems primarily rely on single-dimensional image data, while current deep-hashing methods are capable of learning complex feature representations. However, retrieval accuracy and efficiency are hindered by diverse modalities and limited sample sizes.
View Article and Find Full Text PDFIEEE Trans Neural Netw Learn Syst
July 2024
Cross-modal hashing encodes different modalities of multimodal data into low-dimensional Hamming space for fast cross-modal retrieval. In multi-label cross-modal retrieval, multimodal data are often annotated with multiple labels, and some labels, e.g.
View Article and Find Full Text PDFIEEE Trans Pattern Anal Mach Intell
December 2024
Cross-modal hashing (CMH) has attracted considerable attention in recent years. Almost all existing CMH methods primarily focus on reducing the modality gap and semantic gap, i.e.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!