Home > Research > Publications & Outputs > TUCH

Associated organisational unit

Electronic data

  • ijcai2017_submission_xinzhao

    Accepted author manuscript, 1.03 MB, PDF document

    Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License

Links

Text available via DOI:

View graph of relations

TUCH: Turning Cross-view Hashing into Single-view Hashing via Generative Adversarial Nets

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date19/08/2017
Host publicationProceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence
EditorsCarles Sierra
Place of PublicationMelbourne
PublisherIJCAI
Pages3511-3517
Number of pages7
ISBN (electronic)9780999241103
<mark>Original language</mark>English
EventIJCAI17 -
Duration: 21/08/201725/08/2017

Conference

ConferenceIJCAI17
Period21/08/1725/08/17

Conference

ConferenceIJCAI17
Period21/08/1725/08/17

Abstract

Cross-view retrieval, which focuses on searching images as response to text queries or vice versa, has received increasing attention recently. Cross-view hashing is to efficiently solve the cross-view retrieval problem with binary hash codes. Most existing works on cross-view hashing exploit multi-view embedding method to tackle this problem, which inevitably causes the information loss in both image and text domains. Inspired by the Generative Adversarial Nets (GANs), this paper presents a new model that is able to Turn Cross-view Hashing into single-view hashing (TUCH), thus enabling the information of image to be preserved as much as possible. TUCH is a novel deep architecture that integrates a language model network T for text feature extraction, a generator network G to generate fake images from text feature and a hashing network H for learning hashing functions to generate compact binary codes. Our architecture effectively unifies joint generative adversarial learning and cross-view hashing. Extensive empirical evidence shows that our TUCH approach achieves state-of-the-art results, especially on text to image retrieval, based on image-sentences datasets, i.e. standard IAPRTC-12 and large-scale Microsoft COCO.