{"publisher":"Springer Nature","day":"01","publication":"Multimedia Tools and Applications","author":[{"last_name":"Kang","first_name":"Peipei","full_name":"Kang, Peipei"},{"full_name":"Lin, Zehang","last_name":"Lin","first_name":"Zehang"},{"last_name":"Yang","first_name":"Zhenguo","full_name":"Yang, Zhenguo"},{"full_name":"Bronstein, Alexander","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","first_name":"Alexander","orcid":"0000-0001-9699-8730","last_name":"Bronstein"},{"full_name":"Li, Qing","last_name":"Li","first_name":"Qing"},{"full_name":"Liu, Wenyin","last_name":"Liu","first_name":"Wenyin"}],"language":[{"iso":"eng"}],"month":"05","doi":"10.1007/s11042-022-12187-6","article_type":"original","date_published":"2022-05-01T00:00:00Z","status":"public","citation":{"apa":"Kang, P., Lin, Z., Yang, Z., Bronstein, A. M., Li, Q., & Liu, W. (2022). Deep fused two-step cross-modal hashing with multiple semantic supervision. Multimedia Tools and Applications. Springer Nature. https://doi.org/10.1007/s11042-022-12187-6","mla":"Kang, Peipei, et al. “Deep Fused Two-Step Cross-Modal Hashing with Multiple Semantic Supervision.” Multimedia Tools and Applications, vol. 81, no. 11, Springer Nature, 2022, pp. 15653–70, doi:10.1007/s11042-022-12187-6.","ista":"Kang P, Lin Z, Yang Z, Bronstein AM, Li Q, Liu W. 2022. Deep fused two-step cross-modal hashing with multiple semantic supervision. Multimedia Tools and Applications. 81(11), 15653–15670.","ieee":"P. Kang, Z. Lin, Z. Yang, A. M. Bronstein, Q. Li, and W. Liu, “Deep fused two-step cross-modal hashing with multiple semantic supervision,” Multimedia Tools and Applications, vol. 81, no. 11. Springer Nature, pp. 15653–15670, 2022.","short":"P. Kang, Z. Lin, Z. Yang, A.M. Bronstein, Q. Li, W. Liu, Multimedia Tools and Applications 81 (2022) 15653–15670.","chicago":"Kang, Peipei, Zehang Lin, Zhenguo Yang, Alex M. Bronstein, Qing Li, and Wenyin Liu. “Deep Fused Two-Step Cross-Modal Hashing with Multiple Semantic Supervision.” Multimedia Tools and Applications. Springer Nature, 2022. https://doi.org/10.1007/s11042-022-12187-6.","ama":"Kang P, Lin Z, Yang Z, Bronstein AM, Li Q, Liu W. Deep fused two-step cross-modal hashing with multiple semantic supervision. Multimedia Tools and Applications. 2022;81(11):15653-15670. doi:10.1007/s11042-022-12187-6"},"publication_status":"published","oa_version":"None","title":"Deep fused two-step cross-modal hashing with multiple semantic supervision","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","issue":"11","_id":"18227","date_updated":"2024-10-14T11:10:00Z","article_processing_charge":"No","abstract":[{"text":"Existing cross-modal hashing methods ignore the informative multimodal joint information and cannot fully exploit the semantic labels. In this paper, we propose a deep fused two-step cross-modal hashing (DFTH) framework with multiple semantic supervision. In the first step, DFTH learns unified hash codes for instances by a fusion network. Semantic label and similarity reconstruction have been introduced to acquire binary codes that are informative, discriminative and semantic similarity preserving. In the second step, two modality-specific hash networks are learned under the supervision of common hash codes reconstruction, label reconstruction, and intra-modal and inter-modal semantic similarity reconstruction. The modality-specific hash networks can generate semantic preserving binary codes for out-of-sample queries. To deal with the vanishing gradients of binarization, continuous differentiable tanh is introduced to approximate the discrete sign function, making the networks able to back-propagate by automatic gradient computation. Extensive experiments on MIRFlickr25K and NUS-WIDE show the superiority of DFTH over state-of-the-art methods.","lang":"eng"}],"intvolume":" 81","scopus_import":"1","date_created":"2024-10-08T12:55:04Z","extern":"1","type":"journal_article","volume":81,"page":"15653-15670","quality_controlled":"1","publication_identifier":{"issn":["1380-7501"],"eissn":["1573-7721"]},"year":"2022"}