1 code implementation • 25 Apr 2018 • Jinwei Qi, Yuxin Peng, Yuxin Yuan
First, we propose visual-language relation attention model to explore both fine-grained patches and their relations of different media types.
no code implementations • 14 Oct 2017 • Yuxin Peng, Jinwei Qi, Yuxin Yuan
They can not only exploit cross-modal correlation for learning common representation, but also preserve reconstruction information for capturing semantic consistency within each modality.
1 code implementation • 16 Aug 2017 • Yuxin Peng, Jinwei Qi, Yuxin Yuan
Effectively measuring the similarity between different modalities of data is the key of cross-modal retrieval.
no code implementations • 14 Apr 2017 • Jinwei Qi, Xin Huang, Yuxin Peng
Motivated by the strong ability of deep neural network in feature representation and comparison functions learning, we propose the Unified Network for Cross-media Similarity Metric (UNCSM) to associate cross-media shared representation learning with distance metric in a unified framework.