Highlights Cross-Lingual Word Embeddings

1
Unsupervised Multilingual Word Embeddings Xilun Chen and Claire Cardie {xlchen, cardie}@cs.cornell.edu Multilingual Adversarial Training Multilingual Pseudo- Supervised Refinement Experiments Word Translation Cross-Lingual Word Embeddings Monolingual cat dog swim walk Bilingual cat dog swim walk chat chien nager marcher Multilingual cat dog swim walk chat chien nager marcher 游泳 Supervised Unsupervised Bilingual Mikolov et al. (2013) Zou et al. (2013) & many more… Zhang et al. (2017) Conneau et al. (2017) Artetxe et al. (2017; 2018) Multilingual Ammar et al. (2016) Duong et al. (2017) This work! en-de en-fr en-es en-it en-pt de-fr de-es de-it de-pt fr-es fr-it fr-pt es-it es-pt it-pt Supervised methods with cross-lingual supervision Sup-BWE-Direct 73.5 81.1 81.4 77.3 79.9 73.3 67.7 69.5 59.1 82.6 83.2 78.1 83.5 87.3 81.0 Unsupervised methods without cross-lingual supervision BWE-Pivot 74.0 82.3 81.7 77.0 80.7 71.9 66.1 68.0 57.4 81.1 79.7 74.7 81.9 85.0 78.9 BWE-Direct 74.0 82.3 81.7 77.0 80.7 73.0 65.7 66.5 58.5 83.1 83.0 77.9 83.3 87.3 80.5 MAT+MPSR 74.8 82.4 82.5 78.8 81.5 76.7 69.6 72.0 63.2 83.9 83.5 79.3 84.5 87.8 82.3 de-en fr-en es-en it-en pt-en fr-de es-de it-de pt-de es-fr it-fr pt-fr it-es pt-es pt-it Supervised methods with cross-lingual supervision Sup-BWE-Direct 72.4 82.4 82.9 76.9 80.3 69.5 68.3 67.5 63.7 85.8 87.1 84.3 87.3 91.5 81.1 Unsupervised methods without cross-lingual supervision BWE-Pivot 72.2 82.1 83.3 77.7 80.1 68.1 67.9 66.1 63.1 84.7 86.5 82.6 85.8 91.3 79.2 BWE-Direct 72.2 82.1 83.3 77.7 80.1 69.7 68.8 62.5 60.5 86 87.6 83.9 87.7 92.1 80.6 MAT+MPSR 72.9 81.8 83.7 77.4 79.9 71.2 69.0 69.5 65.7 86.9 88.1 86.3 88.2 92.7 82.6 (a) Detailed Results Training Cost Single Source Single Target #BWEs time en-xx de-xx fr-xx es-xx it-xx pt-xx xx-en xx-de xx-fr xx-es xx-it xx-pt Overall Supervised methods with cross-lingual supervision Sup-BWE-Direct N (N -1) 4h 78.6 68.4 79.2 81.6 80.0 80.2 79.0 68.5 82.3 82.1 78.9 77.1 78.0 Unsupervised methods without cross-lingual supervision BWE-Pivot 2(N -1) 8h 79.1 67.1 77.1 80.6 79.0 79.3 79.1 67.8 81.6 81.2 77.2 75.3 77.0 BWE-Direct N (N -1) 23h 79.1 67.2 79.2 81.7 79.2 79.4 79.1 67.1 82.6 82.1 78.1 77.0 77.6 MAT+MPSR N -1 5h 80.0 70.9 79.9 82.4 81.1 81.4 79.1 70.0 84.1 83.4 80.3 78.8 79.3 (b) Summarized Results Table 1: Multilingual Word Translation Results for English, German, French, Spanish, Italian and Portuguese. The reported numbers are precision@1 in percentage. All systems use the nearest neighbor under the CSLS distance for predicting the translation of a certain word. Method Embeddings from lang i lang i Encoder lang j Decoder lang j Discriminator Shared Embedding Space Embeddings from lang j Forward and backward passes when training M Forward and backward passes when training D D j M i M > j J D j J M i lang i lang i lang j lang j lang j Algorithm 1 Multilingual Adversarial Training Require: Vocabulary V i for each language lang i 2 L . Hy- perparameter k 2 N. 1: repeat 2: . D iterations 3: for diter =1 to k do 4: loss d =0 5: for all lang j 2 L do 6: Select at random lang i 2 L 7: Sample a batch of words x i V i 8: Sample a batch of words x j V j 9: ˆ x t = M i (x i ) . encode to T 10: ˆ x j = M > j x t ) . decode to S j 11: y j = D j (x j ) . real vectors 12: ˆ y j = D j x j ) . converted vectors 13: loss d += L d (1,y j )+ L d (0, ˆ y j ) 14: Update all D parameters to minimize loss d 15: . M iteration 16: loss =0 17: for all lang i 2 L do 18: Select at random lang j 2 L 19: Sample a batch of words x i V i 20: ˆ x t = M i (x i ) . encode to T 21: ˆ x j = M > j x t ) . decode to S j 22: ˆ y j = D j x j ) 23: loss += L d (1, ˆ y j ) 24: Update all M parameters to minimize loss 25: orthogonalize(M) . see §3.3 26: until convergence Algorithm 2 Multilingual Pseudo-Supervised Re- finement Require: A set of (pseudo-)supervised lexica of word pairs between each pair of languages Lex(lang i , lang j ). 1: repeat 2: loss =0 3: for all lang i 2 L do 4: Select at random lang j 2 L 5: Sample (x i ,x j ) Lex(lang i , lang j ) 6: t i = M i (x i ) . encode x i 7: t j = M j (x j ) . encode x j 8: loss += L r (t i ,t j ) . refinement loss 9: Update all M parameters to minimize loss 10: orthogonalize(M) . see §3.3 11: until convergence en-de en-es de-es en-it de-it es-it en-fa de-fa es-fa it-fa Average Supervised methods with cross-lingual supervision Luminoso .769 .772 .735 .787 .747 .767 .595 .587 .634 .606 .700 NASARI .594 .630 .548 .647 .557 .592 .492 .452 .466 .475 .545 Unsupervised methods without cross-lingual supervision BWE-Pivot .709 .711 .703 .709 .682 .721 .672 .655 .701 .688 .695 BWE-Direct .709 .711 .703 .709 .675 .726 .672 .662 .714 .695 .698 MAT+MPSR .711 .712 .708 .709 .684 .730 .680 .674 .720 .709 .704 Table 2: Results for the SemEval-2017 Cross-Lingual Word Similarity task. Spearman’s is reported. Lumi- noso (Speer and Lowry-Duda, 2017) and NASARI (Camacho-Collados et al., 2016) are the two top-performing systems for SemEval-2017 that reported results on all language pairs. N languages, each with trained monolingual embeddings as input Learn N-1 orthogonal matrices to map all languages into the same space Explicitly model the interaction between all pairs of languages Despite exploiting O(N 2 ) language pairs, our method scales linearly with N Step 1: Multilingual Adversarial Training (MAT) Step 2: Multilingual Pseudo-Supervised Refinement (MPSR) MAT does a good job for more frequent words but may produce noisier signals for rare words. MPSR anchors on the more accurately predicted relations between frequent words to improve performance on full vocabulary. Lex(lang i , lang j ) is constructed using mutual nearest neighbors among 15k most frequent words 6 languages: English, German, French, Spanish, Italian, Portuguese Word translation retrieved as nearest neighbors in embeddings space Baselines BWE-Pivot: Map each language independently from and to English In total 2(N-1) MUSE (Conneau et al., 2017) BWEs Use English as a pivot for word translation: e.g. de -> en -> fr BWE-Direct: Learn N(N-1) MUSE BWEs for each language pair Sup-BWE-Direct: Learn N(N-1) Supervised BWEs for each language pair Each pair uses 5k labeled word pairs for training Cross-Lingual Word Similarity Dataset from SemEval-2017 Shared Task Evaluates how well the similarity in the cross-lingual embedding space corresponds to a human-annotated semantic similarity score Luminoso and NASARI have access to EuroParl and OpenSubtitles2016 parallel corpora Website: http://www.cs.cornell.edu/~xlchen/ Code: https://github.com/ccsasuke/umwe Cross-Lingual Supervision Parallel Corpus Bilingual Lexicon The cat sleeps on the couch. 猫睡在沙发上。 cat Training Cost #BWEs time overall Supervised methods Sup-MUSE-Direct 30 4h 78.0 Unsupervised methods MUSE-Pivot 10 8h 77.0 MUSE-Direct 30 23h 77.6 Ours 5 5h 79.3 Word Translation Accuracy on 6 languages (30 pairs) TL;DR: A multilingual generalization of the Facebook MUSE (Conneau et al., 2017) embeddings. Try our method out if you are using MUSE to map multiple languages into a single space! Highlights French Italian English BWE-Pivot: O(N) BWEs French Italian Spanish BWE-Direct: O(N 2 ) BWEs Existing Methods Learning Unsupervised Multilingual Embeddings

Transcript of Highlights Cross-Lingual Word Embeddings

Page 1: Highlights Cross-Lingual Word Embeddings

Unsupervised Multilingual Word EmbeddingsXilun Chen and Claire Cardie

{xlchen, cardie}@cs.cornell.edu

Multilingual Adversarial Training Multilingual Pseudo- Supervised Refinement

Experiments

Word Translation

Cross-Lingual Word Embeddings

Monolingual

cat

dog

swim

walk

Bilingual

cat

dog

swim

walkchat

chien

nager

marcher

Multilingual

cat

dog

swim

walkchat

chien

nager

marcher

游泳

⾛走

Supervised Unsupervised

BilingualMikolov et al. (2013)

Zou et al. (2013)& many more…

Zhang et al. (2017)Conneau et al. (2017)

Artetxe et al. (2017; 2018)

Multilingual Ammar et al. (2016)Duong et al. (2017) This work!

en-de en-fr en-es en-it en-pt de-fr de-es de-it de-pt fr-es fr-it fr-pt es-it es-pt it-pt

Supervised methods with cross-lingual supervisionSup-BWE-Direct 73.5 81.1 81.4 77.3 79.9 73.3 67.7 69.5 59.1 82.6 83.2 78.1 83.5 87.3 81.0

Unsupervised methods without cross-lingual supervisionBWE-Pivot 74.0 82.3 81.7 77.0 80.7 71.9 66.1 68.0 57.4 81.1 79.7 74.7 81.9 85.0 78.9BWE-Direct 74.0 82.3 81.7 77.0 80.7 73.0 65.7 66.5 58.5 83.1 83.0 77.9 83.3 87.3 80.5

MAT+MPSR 74.8 82.4 82.5 78.8 81.5 76.7 69.6 72.0 63.2 83.9 83.5 79.3 84.5 87.8 82.3

de-en fr-en es-en it-en pt-en fr-de es-de it-de pt-de es-fr it-fr pt-fr it-es pt-es pt-it

Supervised methods with cross-lingual supervisionSup-BWE-Direct 72.4 82.4 82.9 76.9 80.3 69.5 68.3 67.5 63.7 85.8 87.1 84.3 87.3 91.5 81.1

Unsupervised methods without cross-lingual supervisionBWE-Pivot 72.2 82.1 83.3 77.7 80.1 68.1 67.9 66.1 63.1 84.7 86.5 82.6 85.8 91.3 79.2BWE-Direct 72.2 82.1 83.3 77.7 80.1 69.7 68.8 62.5 60.5 86 87.6 83.9 87.7 92.1 80.6

MAT+MPSR 72.9 81.8 83.7 77.4 79.9 71.2 69.0 69.5 65.7 86.9 88.1 86.3 88.2 92.7 82.6

(a) Detailed Results

Training Cost Single Source Single Target

#BWEs time en-xx de-xx fr-xx es-xx it-xx pt-xx xx-en xx-de xx-fr xx-es xx-it xx-pt Overall

Supervised methods with cross-lingual supervisionSup-BWE-Direct N(N�1) 4h 78.6 68.4 79.2 81.6 80.0 80.2 79.0 68.5 82.3 82.1 78.9 77.1 78.0

Unsupervised methods without cross-lingual supervisionBWE-Pivot 2(N�1) 8h 79.1 67.1 77.1 80.6 79.0 79.3 79.1 67.8 81.6 81.2 77.2 75.3 77.0BWE-Direct N(N�1) 23h 79.1 67.2 79.2 81.7 79.2 79.4 79.1 67.1 82.6 82.1 78.1 77.0 77.6

MAT+MPSR N�1 5h 80.0 70.9 79.9 82.4 81.1 81.4 79.1 70.0 84.1 83.4 80.3 78.8 79.3

(b) Summarized Results

Table 1: Multilingual Word Translation Results for English, German, French, Spanish, Italian and Portuguese. Thereported numbers are precision@1 in percentage. All systems use the nearest neighbor under the CSLS distancefor predicting the translation of a certain word.

similar languages will result in reduced accuracy.Our MAT+MPSR method, however, overcomes thisdisadvantage of BWE-Pivot and achieves the bestperformance on all these pairs through an explicitmultilingual learning mechanism without increas-ing the computational cost.

Furthermore, our method also beats the BWE-Direct approach, which supports our second hy-pothesis that utilizing knowledge from languagesbeyond the pair itself could improve performance.For instance, there are a few pairs where BWE-Pivot outperforms BWE-Direct, such as de-it, it-de and pt-de, even though it goes through a thirdlanguage (English) in BWE-Pivot. This mightsuggest that for some less similar language pairs,leveraging a third language as a bridge could insome cases work better than only relying on thelanguage pair itself. German is involved in all

these language pairs where BWE-Pivot outper-forms than BWE-Direct, which is potentially dueto the similarity between German and the pivotlanguage English. We speculate that if choosinga different pivot language, there might be otherpairs that could benefit. This observation servesas a possible explanation of the superior perfor-mance of our multilingual method over BWE-Direct, since our method utilizes knowledge fromall languages during training.

4.2 Cross-Lingual Word Similarity

In this section, we evaluate the quality ofour MWEs on the cross-lingual word similarity(CLWS) task, which assesses how well the sim-ilarity in the cross-lingual embedding space cor-responds to a human-annotated semantic similar-ity score. The high-quality CLWS dataset fromSemEval-2017 (Camacho-Collados et al., 2017) is

Method

Embeddings from langi

langi Encoder

langj Decoder

langj Discriminator

Shared Embedding Space

Embeddings from langj

Forward and backward passes when training MForward and backward passes when training D

Dj

Mi

M>j

JDj JMi

langi

langi

langj

langj

langj

Algorithm 1 Multilingual Adversarial TrainingRequire: Vocabulary Vi for each language langi 2 L . Hy-

perparameter k 2 N.1: repeat

2: . D iterations3: for diter = 1 to k do

4: lossd = 05: for all langj 2 L do

6: Select at random langi 2 L7: Sample a batch of words xi ⇠ Vi

8: Sample a batch of words xj ⇠ Vj

9: xt = Mi(xi) . encode to T10: xj = M>

j (xt) . decode to Sj

11: yj = Dj(xj) . real vectors12: yj = Dj(xj) . converted vectors13: lossd += Ld(1, yj) + Ld(0, yj)

14: Update all D parameters to minimize lossd15: . M iteration16: loss = 017: for all langi 2 L do

18: Select at random langj 2 L19: Sample a batch of words xi ⇠ Vi

20: xt = Mi(xi) . encode to T21: xj = M>

j (xt) . decode to Sj

22: yj = Dj(xj)23: loss += Ld(1, yj)

24: Update all M parameters to minimize loss25: orthogonalize(M) . see §3.326: until convergence

same space as a selected target language.In order to learn a multilingual embedding

space without supervision, we employ a seriesof language discriminators Dl, one for each lan-guage l 2 L . Each Dl is a binary classifier witha sigmoid layer on top, and is trained to identifyhow likely a given vector is from Sl, the embed-ding space of language l. On the other hand, totrain the mappings, we convert a vector from a ran-dom language langi to another random languagelangj (via the target space T first). The objectiveof the mappings is to confuse Dj , the language dis-criminator for langj , so the mappings are updatedin a way that Dj cannot differentiate the convertedvectors from the real vectors in Sj . This multilin-gual objective enables us to explicitly exploit therelations between all language pairs during train-ing, leading to improved performance.

Formally, for any language langj , the objectivethat Dj is minimizing is:

JDj = Ei⇠L

Exi⇠Sixj⇠Sj

rLd (1, Dj(xj))+

Ld

⇣0, Dj(M>

j Mixi)⌘z (1)

where Ld(y, y) is the loss function of D, whichis chosen as the cross entropy loss in practice. y

is the language label with y = 1 indicates a realembedding from that language.

Furthermore, the objective of Mi for langi is:

JMi = Ej⇠L

Exi⇠Sixj⇠Sj

Ld

⇣1, Dj(M>

j Mixi)⌘

(2)

where Mi strives to make Dj believe that a con-verted vector to langj is instead real. This adver-sarial relation between M and D stimulates Mto learn a shared multilingual embedding space bymaking the converted vectors look as authentic aspossible so that D cannot predict whether a vectoris a genuine embedding from a certain language orconverted from another language via M.

In addition, we allow langi and langj to bethe same language in (1) and (2). In this case,we are encoding a language to T and back toitself, essentially forming an adversarial autoen-coder (Makhzani et al., 2015), which is reportedto improve the model performance (Zhang et al.,2017). Finally, on Line 5 and 17 in Algorithm 1, afor loop is used instead of random sampling. Thisis to ensure that in each step, every discrimina-tor (or mapping) is getting updated at least once,so that we do not need to increase the number oftraining iterations when adding more languages.Computationally, when compared to the BWE-Pivot and BWE-Direct baselines, one step of MATtraining costs similarly to N BWE training steps,and in practice we train MAT for the same num-ber of iterations as training the baselines. There-fore, MAT training scales linearly with the num-ber of languages similar to BWE-Pivot (instead ofquadratically as in BWE-Direct).

3.2 Multilingual Pseudo-Supervised

Refinement

Using MAT, we are able to obtain UMWEs withreasonable quality, but they do not yet achievestate-of-the-art performance. Previous researchon learning unsupervised BWEs (Lample et al.,2018b) observes that the embeddings obtainedfrom adversarial training do a good job aligningthe frequent words between two languages, butperformance degrades when considering the fullvocabulary. They hence propose to use an iter-ative refinement method (Artetxe et al., 2017) torepeatedly refine the embeddings obtained fromthe adversarial training. The idea is that we cananchor on the more accurately predicted relationsbetween frequent words to improve the mappingslearned by adversarial training.

Algorithm 2 Multilingual Pseudo-Supervised Re-finementRequire: A set of (pseudo-)supervised lexica of word pairs

between each pair of languages Lex(langi, langj).1: repeat

2: loss = 03: for all langi 2 L do

4: Select at random langj 2 L5: Sample (xi, xj) ⇠ Lex(langi, langj)6: ti = Mi(xi) . encode xi

7: tj = Mj(xj) . encode xj

8: loss += Lr(ti, tj) . refinement loss9: Update all M parameters to minimize loss

10: orthogonalize(M) . see §3.311: until convergence

When learning MWEs, however, it is desirableto go beyond aligning each language with the tar-get space individually, and instead utilize the re-lations between all languages as we did in MAT.Therefore, we in this section propose a general-ization of the existing refinement methods to in-corporate a multilingual objective.

In particular, MAT can produce an approxi-mately aligned embedding space. As mentionedearlier, however, the training signals from D forrare words are noisier and may lead to worseperformance. Thus, the idea of MultilingualPseudo-Supervised Refinement (MPSR) is to in-duce a dictionary of highly confident word pairsfor every language pair, used as pseudo supervi-sion to improve the embeddings learned by MAT.For a specific language pair (langi, langj), thepseudo-supervised lexicon Lex(langi, langj) isconstructed from mutual nearest neighbors be-tween MiEi and MjEj , among the most frequent15k words of both languages.

With the constructed lexica, the MPSR objectiveis:

Jr = E(i,j)⇠L 2

E(xi,xj)⇠Lex(i,j)

Lr(Mixi, Mjxj)

(3)where Lr(x, x) is the loss function for MPSR, forwhich we use the mean square loss. The MPSRtraining is depicted in Algorithm 2.

Cross-Lingual Similarity Scaling (CSLS)

When constructing the pseudo-supervised lexica,a distance metric between embeddings is neededto compute nearest neighbors. Standard distancemetrics such as the Euclidean distance or cosinesimilarity, however, can lead to the hubnessproblem in high-dimensional spaces when usedto calculate nearest neighbors (Radovanovic

et al., 2010; Dinu and Baroni, 2015). Namely,some words are very likely to be the nearestneighbors of many others (hubs), while othersare not the nearest neighbor of any word. Thisproblem is addressed in the literature by designingalternative distance metrics, such as the invertedsoftmax (Smith et al., 2017) or the CSLS (Lampleet al., 2018b). In this work, we adopt the CSLSsimilarity as a drop-in replacement for cosinesimilarity whenever a distance metric is needed.The CSLS similarity (whose negation is a distancemetric) is calculated as follows:

CSLS(x, y) = 2 cos(x, y)

� 1

n

X

y02NY (x)

cos(x, y0)

� 1

n

X

x02NX(y)

cos(x0, y)

(4)

where NY (x) is the set of n nearest neighbors ofx in the vector space that y comes from: Y ={y1, ..., y|Y |}, and vice versa for NX(y). In prac-tice, we use n = 10.

3.3 Orthogonalization

As mentioned in §3, orthogonal linear mappingsare the preferred choice when learning transforma-tions between the embedding spaces of differentlanguages (Xing et al., 2015; Smith et al., 2017).Therefore, we perform an orthogonalization up-date (Cisse et al., 2017) after each training step toensure that our mappings M are (approximately)orthogonal:

8l : Ml = (1 + �)Ml � �MlM>l Ml

where � is set to 0.001.

3.4 Unsupervised Multilingual Validation

In order to do model selection in the unsupervisedsetting, where no validation set can be used, a sur-rogate validation criterion is required that does notdepend on bilingual data. Previous work showspromising results using such surrogate criteria formodel validation in the bilingual case (Lampleet al., 2018b), and we in this work adopt a vari-ant adapted to our multilingual setting:

V (M, E) = E(i,j)⇠Pij

mean csls(M>j MiEi, Ej)

=X

i 6=j

pij · mean csls(M>j MiEi, Ej)

en-de en-es de-es en-it de-it es-it en-fa de-fa es-fa it-fa Average

Supervised methods with cross-lingual supervisionLuminoso .769 .772 .735 .787 .747 .767 .595 .587 .634 .606 .700NASARI .594 .630 .548 .647 .557 .592 .492 .452 .466 .475 .545

Unsupervised methods without cross-lingual supervisionBWE-Pivot .709 .711 .703 .709 .682 .721 .672 .655 .701 .688 .695BWE-Direct .709 .711 .703 .709 .675 .726 .672 .662 .714 .695 .698

MAT+MPSR .711 .712 .708 .709 .684 .730 .680 .674 .720 .709 .704

Table 2: Results for the SemEval-2017 Cross-Lingual Word Similarity task. Spearman’s ⇢ is reported. Lumi-noso (Speer and Lowry-Duda, 2017) and NASARI (Camacho-Collados et al., 2016) are the two top-performingsystems for SemEval-2017 that reported results on all language pairs.

used for evaluation. The dataset contains wordpairs from any two of the five languages: English,German, Spanish, Italian, and Farsi (Persian), an-notated with semantic similarity scores.

In addition to the BWE-Pivot and BWE-Direct baseline methods, we also include thetwo best-performing systems on SemEval-2017,Luminoso (Speer and Lowry-Duda, 2017) andNASARI (Camacho-Collados et al., 2016) forcomparison. Note that these two methods are su-pervised, and have access to the Europarl3 (for alllanguages but Farsi) and the OpenSubtitles20164

parallel corpora.Table 2 shows the results, where the perfor-

mance of each model is measured by the Spear-man correlation. When compared to the BWE-Pivot and the BWE-Direct baselines, MAT+MPSRcontinues to perform the best on all language pairs.The qualitative findings stay the same as in theword translation task, except the margin is less sig-nificant. This might be because the CLWS task ismuch more lenient compared to the word transla-tion task, where in the latter one needs to correctlyidentify the translation of a word out of hundredsof thousands of words in the vocabulary. In CLWSthough, one can still achieve relatively high corre-lation in spite of minor inaccuracies.

On the other hand, an encouraging result isthat when compared to the state-of-the-art super-vised results, our MAT+MPSRmethod outperformsNASARI by a very large margin, and achievestop-notch overall performance similar to the com-petition winner, Luminoso, without using any bi-texts. A closer examination reveals that our unsu-pervised method lags a few points behind Lumi-

3http://opus.nlpl.eu/Europarl.php

4http://opus.nlpl.eu/

OpenSubtitles2016.php

noso on the European languages wherein the su-pervised methods have access to the large-scalehigh-quality Europarl parallel corpora. It is thelow-resource language, Farsi, that makes our un-supervised method stand out. All of the unsuper-vised methods outperform the supervised systemsfrom SemEval-2017 on language pairs involvingFarsi, which is not covered by the Europarl bitexts.This suggests the advantage of learning unsuper-vised embeddings for lower-resourced languages,where the supervision might be noisy or absent.Furthermore, within the unsupervised methods,MAT+MPSR again performs the best, and attainsa higher margin over the baseline approaches onthe low-resource language pairs, vindicating ourclaim of better multilingual performance.

5 Conclusion

In this work, we propose a fully unsupervisedmodel for learning multilingual word embeddings(MWEs). Although methods exist for learninghigh-quality unsupervised BWEs (Lample et al.,2018b), little work has been done in the unsuper-vised multilingual setting. Previous work reliessolely on a number of unsupervised BWE modelsto generate MWEs (e.g. BWE-Pivot and BWE-Direct), which does not fully leverage the interde-pendencies among all the languages. Therefore,we propose the MAT+MPSR method that explicitlyexploits the relations between all language pairswithout increasing the computational cost. In ourexperiments on multilingual word translation andcross-lingual word similarity (SemEval-2017), weshow that MAT+MPSR outperforms existing unsu-pervised and even supervised models, achievingnew state-of-the-art performance.

For future work, we plan to investigate how our

•N languages, each with trained monolingual embeddings as input

•Learn N-1 orthogonal matrices to map all languages into the same space

•Explicitly model the interaction between all pairs of languages

•Despite exploiting O(N2) language pairs, our method scales linearly with N

•Step 1: Multilingual Adversarial Training (MAT)

•Step 2: Multilingual Pseudo-Supervised Refinement (MPSR)

• MAT does a good job for more frequent words but may produce noisier signals for rare words.

• MPSR anchors on the more accurately predicted relations between frequent words to improve performance on full vocabulary.

• Lex(langi, langj) is constructed using mutual nearest neighbors among 15k most frequent words

• 6 languages: English, German, French, Spanish, Italian, Portuguese

•Word translation retrieved as nearest neighbors in embeddings space

Baselines• BWE-Pivot:

•Map each language independently from and to English

• In total 2(N-1) MUSE (Conneau et al., 2017) BWEs

• Use English as a pivot for word translation: e.g. de -> en -> fr

• BWE-Direct:

• Learn N(N-1) MUSE BWEs for each language pair

• Sup-BWE-Direct:

• Learn N(N-1) Supervised BWEs for each language pair

• Each pair uses 5k labeled word pairs for training

Cross-Lingual Word Similarity

•Dataset from SemEval-2017 Shared Task

• Evaluates how well the similarity in the cross-lingual embedding space corresponds to a human-annotated semantic similarity score

• Luminoso and NASARI have access to EuroParl and OpenSubtitles2016 parallel corpora

Website: http://www.cs.cornell.edu/~xlchen/Code: https://github.com/ccsasuke/umwe

Cross-Lingual Supervision

Parallel Corpus

Bilingual Lexicon

The cat sleeps on the couch. 猫睡在沙发上。

cat 猫

Guillaume Lample, Alexis Conneau, Ludovic De-noyer, and Marc’Aurelio Ranzato. 2018a. Unsu-pervised machine translation using monolingual cor-pora only. In International Conference on LearningRepresentations.

Guillaume Lample, Alexis Conneau, Marc’AurelioRanzato, Ludovic Denoyer, and Herv Jgou. 2018b.Word translation without parallel data. In Interna-tional Conference on Learning Representations.

Alireza Makhzani, Jonathon Shlens, Navdeep Jaitly,Ian Goodfellow, and Brendan Frey. 2015. Adversar-ial autoencoders. arXiv preprint arXiv:1511.05644.

Tomas Mikolov, Quoc V. Le, and Ilya Sutskever. 2013a.Exploiting similarities among languages for ma-chine translation. CoRR, abs/1309.4168.

Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Cor-rado, and Jeffrey Dean. 2013b. Distributed repre-sentations of words and phrases and their composi-tionality. In Proceedings of the 26th InternationalConference on Neural Information Processing Sys-tems - Volume 2, pages 3111–3119, USA. CurranAssociates Inc.

Milos Radovanovic, Alexandros Nanopoulos, and Mir-jana Ivanovic. 2010. Hubs in space: Popular nearestneighbors in high-dimensional data. J. Mach. Learn.Res., 11:2487–2531.

Samuel L. Smith, David H. P. Turban, Steven Hamblin,and Nils Y. Hammerla. 2017. Offline bilingual wordvectors, orthogonal transformations and the invertedsoftmax. In Proceedings of ICLR.

Anders Søgaard, Sebastian Ruder, and Ivan Vulic.2018. On the limitations of unsupervised bilingualdictionary induction. In Proceedings of the 56th An-nual Meeting of the Association for ComputationalLinguistics (Volume 1: Long Papers), pages 778–788. Association for Computational Linguistics.

Robert Speer and Joanna Lowry-Duda. 2017. Con-ceptnet at semeval-2017 task 2: Extending wordembeddings with multilingual relational knowledge.In Proceedings of the 11th International Workshopon Semantic Evaluation, SemEval@ACL 2017, Van-couver, Canada, August 3-4, 2017, pages 85–89.

Joseph Turian, Lev-Arie Ratinov, and Yoshua Bengio.2010. Word representations: A simple and generalmethod for semi-supervised learning. In Proceed-ings of the 48th Annual Meeting of the Associationfor Computational Linguistics, pages 384–394, Up-psala, Sweden. Association for Computational Lin-guistics.

Ivan Vulic and Marie-Francine Moens. 2015. Bilin-gual word embeddings from non-parallel document-aligned data applied to bilingual lexicon induction.In Proceedings of the 53rd Annual Meeting of theAssociation for Computational Linguistics and the7th International Joint Conference on Natural Lan-guage Processing (Volume 2: Short Papers), pages

Training Cost

#BWEs time overall

Supervised methodsSup-MUSE-Direct 30 4h 78.0

Unsupervised methodsMUSE-Pivot 10 8h 77.0MUSE-Direct 30 23h 77.6

Ours 5 5h 79.3

Table 3: Overall Word Translation Accuracy for 6 lan-guages (30 language pairs).

719–725, Beijing, China. Association for Computa-tional Linguistics.

Chao Xing, Dong Wang, Chao Liu, and Yiye Lin.2015. Normalized word embedding and orthog-onal transform for bilingual word translation. InProceedings of the 2015 Conference of the NorthAmerican Chapter of the Association for Computa-tional Linguistics: Human Language Technologies,pages 1006–1011, Denver, Colorado. Associationfor Computational Linguistics.

Meng Zhang, Yang Liu, Huanbo Luan, and MaosongSun. 2017. Adversarial training for unsupervisedbilingual lexicon induction. In Proceedings of the55th Annual Meeting of the Association for Com-putational Linguistics (Volume 1: Long Papers),pages 1959–1970, Vancouver, Canada. Associationfor Computational Linguistics.

Will Y. Zou, Richard Socher, Daniel Cer, and Christo-pher D. Manning. 2013. Bilingual word embeddingsfor phrase-based machine translation. In Proceed-ings of the 2013 Conference on Empirical Methodsin Natural Language Processing, pages 1393–1398,Seattle, Washington, USA. Association for Compu-tational Linguistics.

Word Translation Accuracy on 6 languages (30 pairs)

TL;DR: A multilingual generalization of the Facebook MUSE (Conneau et al., 2017) embeddings. Try our method out if you are using MUSE to map multiple languages into a single space!

Highlights

French Italian

EnglishBWE-Pivot: O(N) BWEs

French Italian

SpanishBWE-Direct: O(N2) BWEs

Existing MethodsLearning Unsupervised Multilingual Embeddings