SOTAVerified

Bridging Languages through Images with Deep Partial Canonical Correlation Analysis

2018-07-01ACL 2018Code Available0· sign in to hype

Guy Rotman, Ivan Vuli{\'c}, Roi Reichart

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We present a deep neural network that leverages images to improve bilingual text embeddings. Relying on bilingual image tags and descriptions, our approach conditions text embedding induction on the shared visual information for both languages, producing highly correlated bilingual embeddings. In particular, we propose a novel model based on Partial Canonical Correlation Analysis (PCCA). While the original PCCA finds linear projections of two views in order to maximize their canonical correlation conditioned on a shared third variable, we introduce a non-linear Deep PCCA (DPCCA) model, and develop a new stochastic iterative algorithm for its optimization. We evaluate PCCA and DPCCA on multilingual word similarity and cross-lingual image description retrieval. Our models outperform a large variety of previous methods, despite not having access to any visual signal during test time inference.

Tasks

Reproductions