Bridging languages through images with deep partial canonical correlation analysis

Authors
Rotman, G 
Vulić, I 
Reichart, R 

Loading...
Thumbnail Image
Type
Conference Object
Change log
Abstract

We present a deep neural network that leverages images to improve bilingual text embeddings. Relying on bilingual image tags and descriptions, our approach conditions text embedding induction on the shared visual information for both languages, producing highly correlated bilingual embeddings. In particular, we propose a novel model based on Partial Canonical Correlation Analysis (PCCA). While the original PCCA finds linear projections of two views in order to maximize their canonical correlation conditioned on a shared third variable, we introduce a non-linear Deep PCCA (DPCCA) model, and develop a new stochastic iterative algorithm for its optimization. We evaluate PCCA and DPCCA on multilingual word similarity and cross-lingual image description retrieval. Our models outperform a large variety of previous methods, despite not having access to any visual signal during test time inference. Our code and data are available at: https://github.com/rotmanguy/DPCCA}

Publication Date
2018
Online Publication Date
Acceptance Date
2018-04-21
Keywords
Journal Title
ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers)
Journal ISSN
Volume Title
1
Publisher
Association for Computational Linguistics
Sponsorship
European Research Council (648909)