Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval
Kang P, Lin Z, Yang Z, Fang X, Bronstein AM, Li Q, Liu W. 2022. Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval. Applied Intelligence. 52, 33–54.
Download
No fulltext has been uploaded. References only!
Journal Article
| Published
| English
Scopus indexed
Author
Kang, Peipei;
Lin, Zehang;
Yang, Zhenguo;
Fang, Xiaozhao;
Bronstein, Alex M.ISTA ;
Li, Qing;
Liu, Wenyin
Abstract
Cross-modal retrieval aims to retrieve related items across different modalities, for example, using an image query to retrieve related text. The existing deep methods ignore both the intra-modal and inter-modal intra-class low-rank structures when fusing various modalities, which decreases the retrieval performance. In this paper, two deep models (denoted as ILCMR and Semi-ILCMR) based on intra-class low-rank regularization are proposed for supervised and semi-supervised cross-modal retrieval, respectively. Specifically, ILCMR integrates the image network and text network into a unified framework to learn a common feature space by imposing three regularization terms to fuse the cross-modal data. First, to align them in the label space, we utilize semantic consistency regularization to convert the data representations to probability distributions over the classes. Second, we introduce an intra-modal low-rank regularization, which encourages the intra-class samples that originate from the same space to be more relevant in the common feature space. Third, an inter-modal low-rank regularization is applied to reduce the cross-modal discrepancy. To enable the low-rank regularization to be optimized using automatic gradients during network back-propagation, we propose the rank-r approximation and specify the explicit gradients for theoretical completeness. In addition to the three regularization terms that rely on label information incorporated by ILCMR, we propose Semi-ILCMR in the semi-supervised regime, which introduces a low-rank constraint before projecting the general representations into the common feature space. Extensive experiments on four public cross-modal datasets demonstrate the superiority of ILCMR and Semi-ILCMR over other state-of-the-art methods.
Publishing Year
Date Published
2022-01-01
Journal Title
Applied Intelligence
Publisher
Springer Nature
Volume
52
Page
33-54
ISSN
eISSN
IST-REx-ID
Cite this
Kang P, Lin Z, Yang Z, et al. Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval. Applied Intelligence. 2022;52:33-54. doi:10.1007/s10489-021-02308-3
Kang, P., Lin, Z., Yang, Z., Fang, X., Bronstein, A. M., Li, Q., & Liu, W. (2022). Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval. Applied Intelligence. Springer Nature. https://doi.org/10.1007/s10489-021-02308-3
Kang, Peipei, Zehang Lin, Zhenguo Yang, Xiaozhao Fang, Alex M. Bronstein, Qing Li, and Wenyin Liu. “Intra-Class Low-Rank Regularization for Supervised and Semi-Supervised Cross-Modal Retrieval.” Applied Intelligence. Springer Nature, 2022. https://doi.org/10.1007/s10489-021-02308-3.
P. Kang et al., “Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval,” Applied Intelligence, vol. 52. Springer Nature, pp. 33–54, 2022.
Kang P, Lin Z, Yang Z, Fang X, Bronstein AM, Li Q, Liu W. 2022. Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval. Applied Intelligence. 52, 33–54.
Kang, Peipei, et al. “Intra-Class Low-Rank Regularization for Supervised and Semi-Supervised Cross-Modal Retrieval.” Applied Intelligence, vol. 52, Springer Nature, 2022, pp. 33–54, doi:10.1007/s10489-021-02308-3.