The Secret is in the Spectra: Predicting Cross-lingual Task Performance with Spectral Similarity Measures
Abstract
Performance in cross-lingual NLP tasks is impacted by the (dis)similarity of languages at hand: e.g., previous work has suggested there is a connection between the expected success of bilingual lexicon induction (BLI) and the assumption of (approximate) isomorphism between monolingual embedding spaces. In this work we present a large-scale study focused on the correlations between monolingual embedding space similarity and task performance, covering thousands of language pairs and four different tasks: BLI, parsing, POS tagging and MT. We hypothesize that statistics of the spectrum of each monolingual embedding space indicate how well they can be aligned. We then introduce several isomorphism measures between two embedding spaces, based on the relevant statistics of their individual spectra. We empirically show that 1) language similarity scores derived from such spectral isomorphism measures are strongly associated with performance observed in different cross-lingual tasks, and 2) our spectral-based measures consistently outperform previous standard isomorphism measures, while being computationally more tractable and easier to interpret. Finally, our measures capture complementary information to typologically driven language distance measures, and the combination of measures from the two families yields even higher task performance correlations.
1 Introduction
The effectiveness of joint multilingual modeling and cross-lingual transfer in cross-lingual NLP is critically impacted by the actual languages in consideration Bender (2011); Ponti et al. (2019). Characterizing, measuring, and understanding this cross-language variation is often the first step towards the development of more robust multilingually applicable NLP technology O’Horan et al. (2016); Bjerva et al. (2019); Ponti et al. (2019). For instance, selecting suitable source languages is a prerequisite for successful cross-lingual transfer of dependency parsers or POS taggers Naseem et al. (2012); Ponti et al. (2018); de Lhoneux et al. (2018). In another example, with all other factors kept similar (e.g., training data size, domain similarity), the quality of machine translation also depends heavily on the properties and language proximity of the actual language pair Kudugunta et al. (2019).
In this work, we contribute to this research endeavor by proposing a suite of spectral-based measures that capture the degree of isomorphism Søgaard et al. (2018) between the monolingual embedding spaces of two languages. Our main hypothesis is that the potential to align two embedding spaces and learn transfer functions can be estimated through the differences between the monolingual embeddings’ spectra. We therefore discuss representative statistics of the spectrum of an embedding space (i.e., the set of the singular values of the embedding matrix), such as its condition number or its sorted list of singular values. We then derive measures for the isomorphism between two embedding spaces based on these statistics.
To validate our hypothesis, we perform an extensive empirical evaluation with a range of cross-lingual NLP tasks. This analysis reveals that our proposed spectrum-based isomorphism measures better correlate and explain greater variance than previous isomorphism measures Søgaard et al. (2018); Patra et al. (2019). In addition, our measures also outperform standard approaches based on linguistic information Littell et al. (2017),
The first part of our empirical analysis targets bilingual lexicon induction (BLI), a cross-lingual task that received plenty of attention, in particular as a case study to investigate the impact of cross-language variation on task performance Søgaard et al. (2018); Artetxe et al. (2018). Its popularity stems from its simple task formulation and reduced resource requirements, which makes it widely applicable across a large number of language pairs Ruder et al. (2019b).
Prior work has empirically verified that for some language pairs BLI performs remarkably well, and for others rather poorly Søgaard et al. (2018); Vulić et al. (2019). It attempted to explain this variance in performance by grounding it in the differences between the monolingual embedding spaces themselves. These studies introduced the notion of approximate isomorphism, and argued that it is easier to learn a mapping function Mikolov et al. (2013); Ruder et al. (2019b) between language pairs whose embeddings are approximately isomorphic, than between languages pairs without this property Barone (2016); Søgaard et al. (2018). Subsequently, novel methods to quantify the degree of isomorphism were proposed, and were shown to significantly correlate with BLI scores Zhang et al. (2017); Søgaard et al. (2018); Patra et al. (2019).
In this work, we report much higher correlations with BLI scores than existing isomorphism measures, across a variety of state-of-the-art BLI approaches. While previous work was limited only to coarse-grained analysis with a small number of language pairs (i.e., ), our study is the first large-scale analysis that is focused on the relationship between quantifiable isomorphism and BLI performance. Our analysis covers hundreds of diverse language pairs, focusing on typologically, geographically and phylogenetically distant pairs as well as on similar languages.
We further show that our findings generalize beyond BLI, to cross-lingual transfer in dependency parsing and POS tagging, and we also demonstrate strong correlations with machine translation (MT) performance. Finally, our spectral-based measures can be combined with typologically driven language distance measures to achieve further correlation improvements. This indicates the complementary nature of the implicit knowledge coded in continuous semantic spaces (and captured by our spectral measures) and the discrete linguistic information from typological databases (captured by the typologically driven measures).
2 Quantifying Isomorphism with Spectral Statistics
Following the distributional hypothesis Harris (1954); Firth (1957), word embedding models learn the meaning of words according to their co-occurrence patterns. Hence, the word embedding space of a language whose words are used in diverse contexts is intuitively expected to encode richer information and greater variance than the word embedding space of a language with more restricting word usage patterns. The difference between two monolingual embedding spaces may also result from other reasons, such as the difference between the training corpora on which the embedding induction algorithm is trained, and the degree to which this algorithm accounts for the linguistic properties of each of the languages.
While the exact combination of factors that govern the difference between the embedding spaces of different languages is hard to figure, this difference is likely to be indicative of the quality of cross-lingual transfer. This is particularly true when the embedding spaces are used by cross-lingual transfer algorithms. Our core hypothesis is that the difference between two monolingual spaces can be quantified by spectral statistics of the two spaces.
2.1 Spectrum Statistics
Given a -dimensional embedding matrix , we perform Singular Value Decomposition (SVD) and obtain a diagonal matrix whose main diagonal comprises of singular values, , sorted in a descending order.111 is mean-centered, column means have been subtracted and are equal to zero. Our aim is to quantify the difference between two embedding spaces by comparing statistics of their singular values. We next describe such statistics and in §2.2 use them to measure the isomorphism between the spaces.
Condition Number. In numerical analysis, a function’s condition number measures the extent of change of the function’s output value conditioned on a small change in the input Blum (2014). Consider the case of , where and are two embedding spaces mapped via . The condition number, , represents the degree to which small perturbations in the input are amplified in the output . Following Higham et al. (2015), we compute the condition number of an input matrix with singular values as the ratio between its first (largest) and last (smallest) singular values:
(1) |
Why is it a relevant statistic? A smaller condition number denotes a more “stable” matrix that is less sensitive to perturbations. Consequently, learning a transfer function from one embedding space to another is more robust to noise when dealing with spaces with smaller . We thus expect that embedding matrices with high condition numbers might impede the learning of good transfer functions in cross-lingual NLP: A function learnt on an embedding space that is sensitive to small perturbations may not generalize well.
Are small singular values reliable? Small singular values are associated with noise, or with the least important information, and many noise reduction techniques remove them Ford (2015). If the smallest singular value indeed captures noise, this might affect the condition number (Eq. (1)). It is thus crucial to distinguish between “small but significant” and “small and insignificant” singular values. This is what we do below.
Effective Rank. Given sorted singular values, how can we determine the last effective singular value? For a matrix with singular values , the -numerical rank can be defined as: , which means that singular values below a certain threshold are removed. However, this formulation introduces a dependency on the hyper-parameter . To avoid this, Roy and Vetterli (2007) proposed an alternative method that considers the full spectrum of singular values before computing the so-called effective rank of the input matrix X:
(2) |
where is the entropy of the matrix ’s normalized singular value distribution , computed as . , rounded to the smaller integer, yields the index of the last singular value that is considered significant, and is interpreted as the effective dimensionality, or rank, of the matrix . If is the dimensionality of the embedding space , and we assume that the number of word vectors in is typically much larger than , it then holds that (see Roy and Vetterli (2007)):
(3) |
The dimensionality of an embedding space is intuitively assumed to be equal to the dimensionality of its constituent vectors: the matrix rank. Effective rank undermines this assumption: with effective rank matrices of the same ‘initial dimensionality’ can have very different ‘true dimensionalities’ Yin and Shen (2018). Effective rank is used for various problems outside NLP, such as source localization for acoustic Tourbabin and Rafaely (2015) and seismic Leeuwenburgh and Arts (2014) waves, video compression Bhaskaranand and Gibson (2010), and for the evaluation of implicit regularization in neural matrix factorization Arora et al. (2019). We propose to use it to inform and improve the estimation of the condition number.
Effective Condition Number. We replace in Eq. (1) with the singular value at the position of ’s effective rank (see Eq. (2)), and compute the effective condition number as follows:
(4) |
In §5 we empirically validate the quality of the effective condition number in comparison to the standard condition number.
Having defined spectral statistics of an embedding space, we move to define means of comparing two spaces using these statistics.
2.2 Spectral-Based Isomorphism Measures
The statistics described in §2.1 capture properties of a single embedding space, but it is not straightforward how to employ them in order to quantify the similarity between two distinct embedding spaces. In what follows, we introduce isomorphism measures based on the spectral statistics.
Let us assume two embedding matrices and , with their condition numbers, and . We combine the two numbers using the harmonic mean function (hm) to derive an isomorphism measure between two embedding spaces, cond-hm():
(5) |
We similarly define the econd-hm measure over and .
Why harmonic mean? The higher the (effective) condition number of an embedding space, the higher its sensitivity to perturbations (i.e., the performance of transfer functions will be low). We view the condition number as a constraining factor on transferability, but what is the right way to evaluate the ‘transferability potential’ of two spaces via their condition numbers? There are multiple ways to combine two condition numbers, but we have empirically validated (§5) that hm is a robust choice that outperforms some other possibilities (e.g., the arithmetic mean). We hypothesize this is because hm treats large discrepancies between two numbers in a manner that leans towards the smaller one (unlike e.g. arithmetic mean). Two noisy and two stable embedding spaces would have high and low hms, respectively, but a noisy embedding space and a stable one would have an hm that leans towards the stable one.222Constraining factors are usually handled by taking the minimum (or maximum) values; however, this approach would impede the validity of our correlation analysis (see later §4) as it would artificially decrease the variance in only one variable (the language with the lowest measure will be systematically chosen). We refer the reader to Appendix C for an analysis. Our results suggest that embedding spaces with small condition numbers can often tolerate noisy mappings from embedding spaces with high condition numbers, which might result from the improved stability of the former spaces.
Singular Value Gap. In addition to cond-hm and econd-hm, we introduce another measure that empirically quantifies the divergence between the full spectral information of two embedding spaces. This measure quantifies the gap between the singular values obtained from the matrices and sorted in descending order. We define the measure of Singular Value Gap (svg) between two -dimensional spaces and , as the squared Euclidean distance between the corresponding sorted singular values after log transform:
(6) |
where and , are the sorted singular values characterizing the two embedding matrices and . The intuition here is that two embedding spaces with similar singular values at the same index will be more isomorphic and therefore easier to align into a shared space, and enable more effective cross-lingual transfer.
In summary, this section has presented methods that estimate the degree of isomorphism between any given pair of embedding spaces, which may differ in their language, training corpus, embedding induction algorithm or in other factors. While the focus of our empirical analysis (§4, §5) is cross-language learning and transfer, we note that the scope of our methods may be wider, and that they have not been developed only with cross-lingual learning in mind.
3 Related Work and Baselines
We now provide an overview of prior research that focused on two relevant themes: 1) measuring approximate isomorphism between two embedding spaces, and 2) more generally, quantifying the (dis)similarity between languages, going beyond isomorphism measures. The discussed approaches will also be used as the main baselines later in §5.
Measuring Approximate Isomorphism. We focus on two standard isomorphism measures from prior work which are most similar to our work, and use them as our main baselines. The first measure, termed Isospectrality (IS) Søgaard et al. (2018), is based on spectral analysis as well, but of the Laplacian eigenvalues of the nearest neighborhood graphs that originate from the initial embedding spaces and (for further technical details see Appendix A). Søgaard et al. (2018) argue that these eigenvalues are compact representations of the graph Laplacian, and that their comparison reveals the degree of (approximate) isomorphism. Although similar in spirit to our approach, constructing nearest neighborhood graphs (and then analyzing their eigenvalues) removes useful information on the interaction between all vectors from the initial space, which our spectral method retains.
The second measure is the Gromov-Hausdorff distance (GH) introduced by Patra et al. (2019). It measures the maximum distance of a set of points to the nearest point in another set, or in other words the worst case distance between two metric spaces and (for further technical details see again Appendix A). Patra et al. (2019) propose this distance to test how well two language embedding spaces can be aligned under an isometric transformation.
While both IS and GH were reported to have strong correlations with BLI performance in prior work, they have not been evaluated in large-scale experiments before. In fact, the correlations were computed on a very small number of language pairs (IS: 8 pairs, GH: 10 pairs). Further, both measures do not scale well computationally. Therefore, for computational tractability, the scores are computed only on the sub-matrices spanning the sub-spaces of the most frequent subsets from the full embedding spaces (IS: 10k words, GH: 5k words). In this work, we provide full-fledged empirical analyses of the two measures on a much larger number of pairs from diverse languages, and compare them against the spectral-based measures introduced in §2. The fact that the proposed spectral-based methods are grounded in linear algebra theory (cf. §2.1) also arguably provides a more intuitive understanding of their theoretical underpinning than what is currently offered in the relevant prior work.
Measuring Language Similarity. At the same time, distances between language pairs can also be captured through (dis)similarities in their discrete linguistic properties, such as overlap in syntactic features, or proximity along the phylogenetic language tree. The properties are typically hand-crafted, and are extracted from available typological databases such as the World Atlas of Languages (WALS) Dryer and Haspelmath (2013) or URIEL Littell et al. (2017), among others O’Horan et al. (2016); Ponti et al. (2019). Such distances were found useful in guiding and informing cross-lingual transfer tasks Cotterell and Heigold (2017); Agić (2017); Lin et al. (2019); Ponti et al. (2019).
In particular, we compare against three precomputed measures of language distance based on the URIEL typological database Littell et al. (2017). Phylogenetic distance (PHY) is derived from the hypothesized phylogenetic tree of language descent. Typological distance (TYP) is computed based on the overlap in syntactic features of languages from the WALS database Dryer and Haspelmath (2013). Geographic distance (GEO) is obtained from the locations where languages are spoken; see the work of Littell et al. (2017) for more details.
We use these isomorphism measures and linguistic measures as language distance measures. We simply compute language distance between two languages and as , where svg, cond-hm, econd-hm, gh, is, phy, typ, geo. Later in §5 we show that “proxy” language distances originating from the proposed spectral-based isomorphism measures (see §2.2) correlate better with cross-lingual transfer scores across several tasks, than these language distances which are based on discrete linguistic properties. We verify that implicit knowledge coded in continuous embedding spaces and linguistic knowledge explicitly coded in external databases often complement each other.
4 Experimental Setup
The conducted empirical analyses can be divided into two major parts. First, we run large-scale BLI analyses across several hundred language pairs from dozens of languages, comparing the correlation of spectral-based isomorphism measures (§2.2) and all baselines (§3) with performance of a wide spectrum of state-of-the-art BLI methods. Second, we run further correlation analyses with performances in cross-lingual downstream tasks: dependency parsing, POS tagging, and MT. We first provide the details of the experiments that are shared between the two parts, and then provide further specifics of each experimental part.
Monolingual Word Embeddings. For all isomorphism measures (svg, cond-hm, econd-hm, gh and is) and languages in our analyses we use publicly available 300-dim monolingual fastText word embeddings pretrained on Wikipedia with exactly the same default settings (see Bojanowski et al. (2017)), length-normalized and trimmed to the 200k most frequent words.333fasttext.cc/docs/en/pretrained-vectors.html
Isomorphism Measures: Technical Details. For our spectral-based measures, we compute a full SVD decomposition (i.e., no dimensionality reduction) of the embedding space. We compute svg scores for BLI based on the first 40 singular values, which we empirically found to produce slightly better results;444The average gains were around 10%; we note that running svg with the full set of singular values also outperforms the baseline measures. for the other tasks we use all singular values. For is and gh, we replicate the experimental setup from prior work: we compute the is score over the top 10k most frequent words in each monolingual space, while the gh score is computed over the top 5k words from each monolingual space.555github.com/joelmoniz/BLISS
4.1 Bilingual Lexicon Induction
We conduct correlation analyses of the results from previous studies that report BLI scores for a large number of language pairs. On top of that, we complement the existing results from previous research with new results obtained with state-of-the-art BLI methods, applied to additional language pairs.
BLI Setups and Scores. Vulić et al. (2019) ran BLI experiments on 210 language pairs, spanning 15 diverse languages. Their training and test dictionaries (5k and 2k translation pairs) are derived from PanLex Baldwin et al. (2010); Kamholz et al. (2014). We complement the original 210 pairs with additional 210 language pairs of 15 closely related (European) languages using dictionaries extracted from PanLex following the procedure of Vulić et al. (2019). With the additional language set, the aim is to probe if isomorphism measures can also capture more subtle and smaller language differences.666The initial set of Vulić et al. (2019) comprises Bulgarian, Catalan, Esperanto, Estonian, Basque, Finnish, Hebrew, Hungarian, Indonesian, Georgian, Korean, Lithuanian, Norwegian, Thai, Turkish. The additional 210 language pairs are only composed of Germanic, Romance and Slavic languages. For a full list of the languages see Table 4 in the appendix.
We also analyze the BLI results of 108 language pairs from MUSE Conneau et al. (2018). This dataset systematically covers English, with 88 translation pairs that involve English as either the source or target language. Finally, we analyze the available BLI results of Glavaš et al. (2019) (referred to as GTrans) that are based on dictionaries obtained from Google Translate and include 28 language pairs spanning 8 different languages. For the full list of language pairs involved in previous BLI studies, we refer the reader to prior work Conneau et al. (2018); Glavaš et al. (2019); Vulić et al. (2019).
BLI Methods in Comparison. The scores in each BLI setup were computed by several state-of-the-art BLI methods based on cross-lingual word embeddings, briefly described here. 1) Sup is the standard supervised method Artetxe et al. (2016); Smith et al. (2017) that learns a mapping between two embedding spaces based on a training dictionary by solving the orthogonal Procrustes problem Schönemann (1966). 2) Sup+ is another standard supervised method that additionally applies a variety of pre-processing and post-processing steps (e.g., whitening, dewhitening, symmetric re-weighting) before and after learning the mapping matrix, see Artetxe et al. (2018). 3) Unsup is a fully unsupervised method based on the “similarity of monolingual similarities” heuristic to extract the seed dictionary from monolingual data. It then uses an iterative self-learning procedure to improve on the initial noisy dictionary Artetxe et al. (2018). For more technical details on the fully unsupervised model, we refer the reader to prior work Ruder et al. (2019a); Vulić et al. (2019).777The Sup+ and Unsup methods are based on the VecMap framework (github.com/artetxem/vecmap) which showed very competitive and robust BLI performance across a wide range of language pairs in recent comparative analyses Glavaš et al. (2019); Vulić et al. (2019); Doval et al. (2019).
In sum, our analyses are conducted in three BLI setups (PanLex, MUSE, GTrans) and examine three types of state-of-the-art mapping-based methods, both supervised and unsupervised (Sup, Sup+, Unsup). Altogether, these span 556 language pairs, and cover both related and distant languages.888We report all results for each BLI method, dictionary and language pairs in the supplementary material (and also here https://tinyurl.com/skn5cf7). We also report scores with another method, RCSLS Joulin et al. (2018), benchmarked in the GTrans BLI setup (see Table 1). Following prior work Glavaš et al. (2019), our BLI evaluation measure is Mean Reciprocal Rank (MRR). We note that identical findings emerge from running the correlation analyses based on Precision@1 scores in lieu of MRR.
4.2 Downstream Tasks
Following the large-scale nature of our BLI analyses, we run similar correlation analyses on several downstream tasks that comprise a large number of (both similar and distant) language pairs.999For the full list of languages that were analyzed throughout all our experiments see Table 4 in the appendix. We rely on results from a recent study of Lin et al. (2019) that focused on cross-lingual transfer performance in MT, dependency parsing, and POS tagging.101010For full details regarding the models used to compute the scores for each downstream task, we refer the interested reader to the work of Lin et al. (2019) and the accompanying repository: https://github.com/neulab/langrank. We note that scores for each language pair in each task have been produced with the same task architectures.
Machine Translation. Lin et al. (2019) report BLEU scores when translating 54 source languages into English as the target language. We report correlations between the different language distance measures and these 54 BLEU scores.
Dependency Parsing. We base our analysis on the cross-lingual zero-shot parser transfer results of Lin et al. (2019): The standard biaffine dependency parser Dozat and Manning (2017); Dozat et al. (2017) is trained on the training portions of Universal Dependencies (UD) treebanks from 31 languages Nivre et al. (2018), and is then used to parse the test treebank of each language, now used as the target language. We report correlations between the language distance measures and the Labeled Attachment Scores (LAS) for all combinations of 31 languages, resulting in 930 pairs.
POS Tagging. We use POS tagging accuracy scores reported by Lin et al. (2019). These scores span 26 low-resource target languages and 60 source languages which measure the utility of each source language to each of the 26 target languages in POS tagging. We use a sample of 840 language pairs for the correlation analysis, as 16 low-resource target languages and 49 source languages have readily available pretrained fastText vectors.
4.3 Correlation Analyses and Statistical Tests
All scores from isomorphism measures and BLI scores were log-transformed111111This is an order-preserving transformation which is often used when the data distribution is skewed (as was the case with our BLI and isomorphism scores). prior to any correlation computation. We report Pearson’s correlation coefficients in all tasks. This allows us to investigate which of the different individual measures is most important to predict task performance.
Regression Analyses. The individual (i.e., single-variable) analyses are not sufficient to account for the complex interdependencies between the distance measures themselves, and how they interact with task performance when combined. Therefore, we also use standard linear stepwise regression model Hocking (1976); Draper and Smith (1998):
(7) |
Here, task performance is predicted using a set of regressors, (i.e., svg, cond-hm, econd-hm, gh, is, phy, typ, geo), that are added to the model incrementally only if their marginal addition to predicting is statistically significant (). This method is useful for finding variables (i.e., in our case distance measures) with maximal and unique contribution to the explanation of , when the variables themselves are strongly cross-correlated, as in our case.
This model is able to: (a) discern which variables overlap in their information; (b) detect variables that complement each other; and (c) evaluate their joint contribution in predicting task performance. We compute the regression model’s score for all statistically significant variables, and report its square-root, .
Importantly, is not a one-number description of a language, but rather an illustrative quantification of the joint contribution of several different distance measures to the explanation of . Its goal is to investigate potential gains achieved through the combination of several distance measures, as opposed to using a single-best measure. The distance measures that are found statistically significant in the regression analyses are marked by superscripts over (see later in Tables 1 and 2).
5 Analyses and Results
PanLex | MUSE | GTrans | |||||||
Sup* | Sup+ | Unsup | Sup | Sup+ | Unsup | Sup | rcsls | Unsup | |
1. svg | -.74 | -.76 | -.74 | -.60 | -.60 | -.61 | -.59 | -.59 | -.53 |
2. cond-hm | -.66 | -.64 | -.56 | -.58 | -.47 | -.39 | -.46 | -.47 | -.42 |
3. econd-hm | -.79 | -.77 | -.70 | -.46 | -.38 | -.39 | -.74 | -.75 | -.56 |
4. gh | -.41 | -.41 | -.46 | -.43 | -.33 | -.28 | -.47 | -.46 | -.50 |
5. is | -.51 | -.53 | -.49 | -.42 | -.42 | -.34 | -.33 | -.33 | -.36 |
6. phy | -.55 | -.55 | -.41 | -.53 | -.62 | -.34 | -.51 | -.50 | -.52 |
7. typ | -.57 | -.52 | -.38 | -.47 | -.64 | -.44 | -.59 | -.59 | -.43 |
8. geo | -.62 | -.66 | -.62 | -.57 | -.54 | -.35 | -.22 | -.22 | -.35 |
.911,3,6-8 | .911,3,6-8 | .821,3,8 | .691,6 | .747,8 | .611 | .923,6,8 | .933,6,8 | .863,6,8 |
MT | DEP | POS | |
1. svg | -.56 | -.79 | -.52 |
2. cond-hm | -.52 | -.60 | -.41 |
3. econd-hm | -.51 | -.66 | -.44 |
4. gh | -.16 | -.70 | -.52 |
5. is | -.13 | -.67 | -.43 |
6. phy | -.45 | -.56 | -.16 |
7. typ | -.66 | -.40 | -.16 |
8. geo | -.26 | -.58 | -.15 |
.741,7 | .871,6-8 | .591,4,5,7 |
The results are summarized in Tables 1 and 2. The first main finding is that our proposed spectral-based isomorphism measures are strongly correlated with performance across all tasks and settings.121212The negative correlations between svg and econd-hm scores and task performance have a clean interpretation: they are distance/dissimilarity measures, so high scores of those measures indicate less similar languages. In fact, they show the strongest individual correlations with task performance among all isomorphism measures and linguistic distances alike. The only exception is the MT task, where our measures fall short of typ (see Table 2), although we mark that they still hold a strong advantage over the baseline gh and is isomorphism measures that do not seem to capture any useful language similarity properties needed for the MT task.
econd-hm systematically outperforms cond-hm on 2 of 3 BLI datasets and 2 of 3 downstream tasks, validating our assumption that discarding the smallest singular values reduces noise. Additionally, svg shows greater stability across tasks and datasets than econd-hm. A general finding across all tasks is that our spectral measures are the most robust isomorphism measures: they substantially outperform the widely used baselines gh and is.
As stepwise regression discerns between overlapping and complementing variables (see §4.3), another finding indicates that our spectral measures are complemented by linguistically driven language distances. Indeed, their combination achieves very high correlation scores. The results demonstrate this across all tasks and settings (see bottom rows of the tables). For instance, when combining spectral measures with the linguistic distances, the regression model reaches outstanding correlation scores up to on PanLex BLI (Table 1); with 420 language pairs, PanLex is the most comprehensive BLI dataset in our study. In addition, gh and is are not chosen as significant regressors in the stepwise regression model, which indicates that they capture less information than our spectral methods.131313This does not imply gh or is do not capture information that is complementary to discrete linguistic information, only that the implicit knowledge captured by gh or is cannot match that of our spectral-based methods, and their combination with linguistic distances obtained from external knowledge sources. Overall, the regression results support the notion that conceptually different distances (i.e., isomorphism-based versus measures based on linguistic properties) capture different properties of similarities between languages, which has a synergistic effect when they are combined.
Concerning individual tasks, we note that our spectral-based measures outperform the baselines regardless of the underlying BLI method. Further, svg is the most informative distance measure in parsing experiments, and all linguistic distances fall behind isomorphism measures. Combining linguistic distances with svg increases the already high correlation from 0.79 to 0.87 (Table 2, second column). For the POS tagging task, svg and gh are on par, and the combination with is and typ increase their correlation from 0.52 to 0.59 (Table 2, third column). This is the only analysis where baseline methods are found significant.
Additional results and analyses are provided in Appendix B. They further demonstrate that our measures also indicate transfer quality of different target languages for a given source language, and transfer quality of source languages for a given target language, for the tasks discussed in this paper.
6 Further Discussion and Conclusion
This work introduces two spectral-based measures, svg and econd-hm, that excel in predicting performance on a variety of cross-lingual tasks. Both measures leverage information from singular values in different ways: econd-hm uses the ratio between two singular values, and is grounded in linear algebra and numerical analysis Blum (2014); Roy and Vetterli (2007), while svg directly utilizes the full range of singular values. We suspect that the use of the full range of singular values is what makes svg more robust across different tasks and datasets, compared to econd-hm that shows greater variance, as observed in our results above.
While the spectral methods are computed solely on word vectors from Wikipedia, the results in the downstream tasks are computed with different sets of embeddings (e.g., multilingual embeddings for dependency parsing), or the embeddings are learnt during training (for POS tagging and MT). We believe that this discrepancy does not constitute a shortcoming of our analyses, but rather the opposite: spectral-based methods maintain their high correlations in the downstream tasks as well, and this supports the notion that these measures might indeed capture deeper linguistic information than mere similarities between embedding spaces.
Our use of effective rank in improving the condition number (via effective condition number) is also inspired by recent work that aimed to automatically detect true dimensionality of embedding spaces. However, previous work has taken an empirical approach by simply tuning embedding dimensionality to evaluation tasks at hand Wang (2019); Raunak et al. (2019); Carrington et al. (2019). Our intention, on the other hand, is to extract the true embedding dimensionality directly from the embedding space. Another recent study Yin and Shen (2018) employed perturbation analysis to study the robustness of embedding spaces to noise in monolingual settings, and established that it is also related to effective dimensionality of the embedding space. All these inspired us to replace the standard matrix rank with effective rank when computing the condition number, and to introduce the statistic of effective condition number in §2.1.
Our study is also the first to compare language distance measures that are based on discrete linguistic information Littell et al. (2017) with measures of isomorphism (i.e., our spectral-based measures, is, gh), which can also be used as proxy language distance measures. Our findings, suggesting that it is possible to effectively combine these two types of language distance measures, call for further research that will advance our understanding of: 1) what knowledge is captured in monolingual and cross-lingual embedding spaces Gerz et al. (2018); Pires et al. (2019); Artetxe et al. (2020); 2) how that knowledge complements or overlaps with linguistic knowledge compiled into lexical-semantic and typological databases Dryer and Haspelmath (2013); Wichmann et al. (2018); Ponti et al. (2019); and 3) how to use the combined knowledge for more effective transfer in cross-lingual NLP applications Ponti et al. (2018); Eisenschlos et al. (2019).
The differences in embedding spaces of different languages do not only depend on linguistic properties of the languages in consideration, but also on other factors such as the chosen training algorithm, underlying training domain, or training data size and quality Søgaard et al. (2018); Arora et al. (2019); Vulić et al. (2020). In future research we also plan an in-depth study of these factors and their relation to our spectral analysis.
We believe that the main insights from this study will inform and guide different cross-lingual transfer learning methods and scenarios in future work. These might range from choosing source languages for transfer in low-data regimes, over monolingual word vector induction guided by the spectral statistics, even to more effective hyperparameter search.
Acknowledgments
The work of IV and AK is supported by the ERC Consolidator Grant LEXICAL: Lexical Acquisition Across Languages (no 648909) awarded to AK. HD is supported by the Blavatnik Postdoctoral Fellowship Programme. HD would like to thank Dr. Avital Hahamy for her contribution to the initiation of this work.
References
- Agić (2017) Željko Agić. 2017. Cross-lingual parser selection for low-resource languages. In Proceedings of the NoDaLiDa 2017 Workshop on Universal Dependencies (UDW 2017), pages 1–10.
- Arora et al. (2019) Sanjeev Arora, Nadav Cohen, Wei Hu, and Yuping Luo. 2019. Implicit regularization in deep matrix factorization. In Proceedings of NeurIPS, pages 7411–7422.
- Artetxe et al. (2016) Mikel Artetxe, Gorka Labaka, and Eneko Agirre. 2016. Learning principled bilingual mappings of word embeddings while preserving monolingual invariance. In Proceedings of EMNLP, pages 2289–2294.
- Artetxe et al. (2018) Mikel Artetxe, Gorka Labaka, and Eneko Agirre. 2018. A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings. In Proceedings of ACL, pages 789–798.
- Artetxe et al. (2020) Mikel Artetxe, Sebastian Ruder, and Dani Yogatama. 2020. On the cross-lingual transferability of monolingual representations. In Proceedings of ACL.
- Baldwin et al. (2010) Timothy Baldwin, Jonathan Pool, and Susan Colowick. 2010. PanLex and LEXTRACT: Translating all words of all languages of the world. In Proceedings of COLING (Demo Papers), pages 37–40.
- Barone (2016) Antonio Valerio Miceli Barone. 2016. Towards cross-lingual distributed representations without parallel text trained with adversarial autoencoders. In Proceedings of the 1st Workshop on Representation Learning for NLP, pages 121–126.
- Bender (2011) Emily M. Bender. 2011. On achieving and evaluating language-independence in NLP. Linguistic Issues in Language Technology, 6(3):1–26.
- Bhaskaranand and Gibson (2010) Malavika Bhaskaranand and Jerry D Gibson. 2010. Spectral entropy-based quantization matrices for H264/AVC video coding. In 2010 Conference Record of the Forty Fourth Asilomar Conference on Signals, Systems and Computers, pages 421–425.
- Bjerva et al. (2019) Johannes Bjerva, Robert Östling, Maria Han Veiga, Jörg Tiedemann, and Isabelle Augenstein. 2019. What do language representations really represent? Computational Linguistics, 45(2):381–389.
- Blum (2014) Lenore Blum. 2014. Alan Turing and the other theory of computation (expanded), volume 42 of Lecture Notes in Logic, pages 48–69. Cambridge University Press.
- Bojanowski et al. (2017) Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword information. Transactions of the Association for Computational Linguistics, 5:135–146.
- Carrington et al. (2019) Rachel Carrington, Karthik Bharath, and Simon Preston. 2019. Invariance and identifiability issues for word embeddings. In Proceedings of NeurIPS, pages 15114–15123.
- Chazal et al. (2009) Frédéric Chazal, David Cohen-Steiner, Leonidas J Guibas, Facundo Mémoli, and Steve Y Oudot. 2009. Gromov-Hausdorff stable signatures for shapes using persistence. In Computer Graphics Forum, volume 28, pages 1393–1403.
- Conneau et al. (2018) Alexis Conneau, Guillaume Lample, Marc’Aurelio Ranzato, Ludovic Denoyer, and Hervé Jégou. 2018. Word translation without parallel data. In Proceedings of ICLR.
- Cotterell and Heigold (2017) Ryan Cotterell and Georg Heigold. 2017. Cross-lingual character-level neural morphological tagging. In Proceedings of EMNLP, pages 748–759.
- Doval et al. (2019) Yerai Doval, Jose Camacho-Collados, Luis Espinosa-Anke, and Steven Schockaert. 2019. On the robustness of unsupervised and semi-supervised cross-lingual word embedding learning. CoRR, abs/1908.07742.
- Dozat and Manning (2017) Timothy Dozat and Christopher D. Manning. 2017. Deep biaffine attention for neural dependency parsing. In Proceedings of ICLR.
- Dozat et al. (2017) Timothy Dozat, Peng Qi, and Christopher D. Manning. 2017. Stanford’s graph-based neural dependency parser at the CoNLL 2017 shared task. In Proceedings of the CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies, pages 20–30.
- Draper and Smith (1998) Norman R. Draper and Harry Smith. 1998. Applied Regression Analysis, 3rd Edition. John Wiley & Sons.
- Dryer and Haspelmath (2013) Matthew S. Dryer and Martin Haspelmath, editors. 2013. WALS Online. Max Planck Institute for Evolutionary Anthropology, Leipzig.
- Eisenschlos et al. (2019) Julian Eisenschlos, Sebastian Ruder, Piotr Czapla, Marcin Kadras, Sylvain Gugger, and Jeremy Howard. 2019. MultiFiT: Efficient multi-lingual language model fine-tuning. In Proceedings of EMNLP, pages 5702–5707.
- Firth (1957) John Rupert Firth. 1957. A synopsis of linguistic theory, 1930-1955. Studies in Linguistic Analysis.
- Ford (2015) William Ford. 2015. Chapter 15 - the singular value decomposition. In William Ford, editor, Numerical Linear Algebra with Applications, pages 299 – 320. Academic Press, Boston.
- Gerz et al. (2018) Daniela Gerz, Ivan Vulić, Edoardo Maria Ponti, Jason Naradowsky, Roi Reichart, and Anna Korhonen. 2018. Language modeling for morphologically rich languages: Character-aware modeling for word-level prediction. Transactions of the Association for Computational Linguistics, 6:451–465.
- Glavaš et al. (2019) Goran Glavaš, Robert Litschko, Sebastian Ruder, and Ivan Vulić. 2019. How to (properly) evaluate cross-lingual word embeddings: On strong baselines, comparative analyses, and some misconceptions. In Proceedings of ACL, pages 710–721.
- Harris (1954) Zellig S. Harris. 1954. Distributional structure. Word, 10(23):146–162.
- Higham et al. (2015) N.J. Higham, M.R. Dennis, P. Glendinning, P.A. Martin, F. Santosa, and J. Tanner. 2015. The Princeton Companion to Applied Mathematics. Princeton University Press.
- Hocking (1976) Ronald R. Hocking. 1976. The analysis and selection of variables in linear regression. Biometrics, 32(1):1–49.
- Joulin et al. (2018) Armand Joulin, Piotr Bojanowski, Tomas Mikolov, Hervé Jégou, and Edouard Grave. 2018. Loss in translation: Learning bilingual word mapping with a retrieval criterion. In Proceedings of EMNLP, pages 2979–2984.
- Kamholz et al. (2014) David Kamholz, Jonathan Pool, and Susan M. Colowick. 2014. PanLex: Building a resource for panlingual lexical translation. In Proceedings of LREC, pages 3145–3150.
- Kudugunta et al. (2019) Sneha Kudugunta, Ankur Bapna, Isaac Caswell, and Orhan Firat. 2019. Investigating multilingual NMT representations at scale. In Proceedings of EMNLP-IJCNLP, pages 1565–1575.
- Leeuwenburgh and Arts (2014) Olwijn Leeuwenburgh and Rob Arts. 2014. Distance parameterization for efficient seismic history matching with the ensemble kalman filter. Computational Geosciences, 18(3-4):535–548.
- de Lhoneux et al. (2018) Miryam de Lhoneux, Johannes Bjerva, Isabelle Augenstein, and Anders Søgaard. 2018. Parameter sharing between dependency parsers for related languages. In Proceedings of EMNLP, pages 4992–4997.
- Lin et al. (2019) Yu-Hsiang Lin, Chian-Yu Chen, Jean Lee, Zirui Li, Yuyan Zhang, Mengzhou Xia, Shruti Rijhwani, Junxian He, Zhisong Zhang, Xuezhe Ma, Antonios Anastasopoulos, Patrick Littell, and Graham Neubig. 2019. Choosing transfer languages for cross-lingual learning. In Proceedings of ACL, pages 3125–3135.
- Littell et al. (2017) Patrick Littell, David R. Mortensen, Ke Lin, Katherine Kairis, Carlisle Turner, and Lori Levin. 2017. URIEL and lang2vec: Representing languages as typological, geographical, and phylogenetic vectors. In Proceedings of EACL, pages 8–14.
- Mikolov et al. (2013) Tomas Mikolov, Quoc V Le, and Ilya Sutskever. 2013. Exploiting similarities among languages for machine translation. arXiv preprint arXiv:1309.4168.
- Naseem et al. (2012) Tahira Naseem, Regina Barzilay, and Amir Globerson. 2012. Selective sharing for multilingual dependency parsing. In Proceedings of ACL, pages 629–637.
- Nivre et al. (2018) Joakim Nivre, Mitchell Abrams, Željko Agić, Lars Ahrenberg, Lene Antonsen, Katya Aplonova, Maria Jesus Aranzabe, et al. 2018. Universal Dependencies 2.3.
- O’Horan et al. (2016) Helen O’Horan, Yevgeni Berzak, Ivan Vulić, Roi Reichart, and Anna Korhonen. 2016. Survey on the use of typological information in natural language processing. In Proceedings of COLING, pages 1297–1308.
- Patra et al. (2019) Barun Patra, Joel Ruben Antony Moniz, Sarthak Garg, Matthew R. Gormley, and Graham Neubig. 2019. Bilingual lexicon induction with semi-supervision in non-isometric embedding spaces. In Proceedings of ACL, pages 184–193.
- Pires et al. (2019) Telmo Pires, Eva Schlinger, and Dan Garrette. 2019. How multilingual is multilingual BERT? In Proceedings of ACL, pages 4996–5001.
- Ponti et al. (2019) Edoardo Maria Ponti, Helen O’Horan, Yevgeni Berzak, Ivan Vulić, Roi Reichart, Thierry Poibeau, Ekaterina Shutova, and Anna Korhonen. 2019. Modeling language variation and universals: A survey on typological linguistics for natural language processing. Computational Linguistics, 45(3):559–601.
- Ponti et al. (2018) Edoardo Maria Ponti, Roi Reichart, Anna Korhonen, and Ivan Vulić. 2018. Isomorphic transfer of syntactic structures in cross-lingual NLP. In Proceedings of ACL, pages 1531–1542.
- Raunak et al. (2019) Vikas Raunak, Vivek Gupta, and Florian Metze. 2019. Effective dimensionality reduction for word embeddings. In Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019), pages 235–243.
- Roy and Vetterli (2007) Olivier Roy and Martin Vetterli. 2007. The effective rank: A measure of effective dimensionality. In Proceedings of the 15th European Signal Processing Conference, pages 606–610.
- Ruder et al. (2019a) Sebastian Ruder, Anders Søgaard, and Ivan Vulić. 2019a. Unsupervised cross-lingual representation learning. In Proceedings of ACL: Tutorial Abstracts, pages 31–38.
- Ruder et al. (2019b) Sebastian Ruder, Ivan Vulić, and Anders Søgaard. 2019b. A survey of cross-lingual word embedding models. Journal of Artificial Intelligence Research, 65:569–631.
- Schönemann (1966) Peter H. Schönemann. 1966. A generalized solution of the orthogonal Procrustes problem. Psychometrika, 31(1):1–10.
- Smith et al. (2017) Samuel L. Smith, David H.P. Turban, Steven Hamblin, and Nils Y. Hammerla. 2017. Offline bilingual word vectors, orthogonal transformations and the inverted softmax. In Proceedings of ICLR.
- Søgaard et al. (2018) Anders Søgaard, Sebastian Ruder, and Ivan Vulić. 2018. On the limitations of unsupervised bilingual dictionary induction. In Proceedings of ACL, pages 778–788.
- Tourbabin and Rafaely (2015) Vladimir Tourbabin and Boaz Rafaely. 2015. Direction of arrival estimation using microphone array processing for moving humanoid robots. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 23(11):2046–2058.
- Vulić et al. (2019) Ivan Vulić, Goran Glavaš, Roi Reichart, and Anna Korhonen. 2019. Do we really need fully unsupervised cross-lingual embeddings? In Proceedings of EMNLP, pages 4398–4409.
- Vulić et al. (2020) Ivan Vulić, Sebastian Ruder, and Anders Søgaard. 2020. Are all good word vector spaces isomorphic? In Proceedings of EMNLP.
- Wang (2019) Yu Wang. 2019. Single training dimension selection for word embedding with PCA. In Proceedings of EMNLP-IJCNLP, pages 3588–3593.
- Wichmann et al. (2018) Søren Wichmann, André Müller, Viveka Velupillai, Cecil H Brown, Eric W Holman, Pamela Brown, Sebastian Sauppe, Oleg Belyaev, Matthias Urban, Zarina Molochieva, et al. 2018. The ASJP database (version 18).
- Yin and Shen (2018) Zi Yin and Yuanyuan Shen. 2018. On the dimensionality of word embedding. In Proceedings of NeurIPS, pages 887–898.
- Zhang et al. (2017) Meng Zhang, Yang Liu, Huanbo Luan, and Maosong Sun. 2017. Earth mover’s distance minimization for unsupervised bilingual lexicon induction. In Proceedings of EMNLP, pages 1934–1945.
Appendix A IS and GH
Isospectrality (IS) After length-normalizing the vectors, Søgaard et al. (2018) compute the nearest neighbor graphs using a subset of the top most frequent words in each space, and then calculate the Laplacian matrices and of each graph. For , the smallest is then sought such that the sum of its largest eigenvalues is at least 90% of the sum of all its eigenvalues. The same procedure is used to find . They then define . The final IS measure is then the sum of the squared differences of the largest Laplacian eigenvalues: . The lower , the more similar are the graphs and, consequently, the more isomorphic are the two embedding spaces.
Gromov-Hausdorff Distance (GH) It measures the worst case distance between two metric spaces and with a distance function as:
(8) |
It computes the distance between the nearest neighbors that are farthest apart. The Gromov-Hausdorff distance then minimizes this distance over all isometric transforms and :
(9) |
Computing directly is computationally intractable in practice, but it can be tractably approximated by computing the Bottleneck distance between the metric spaces Chazal et al. (2009).
Appendix B Source and Target Selection Analysis
In addition to the correlation analyses in the main text that aggregate all language pairs, some tasks and datasets also support analyses where one language is fixed as a target language (i.e., source-language selection analysis), or as a source language (i.e., target-language selection analysis). Such analyses could inform us on how to choose the right transfer language. That is, given a target language one would like to transfer to, which is the best source language to transfer from, and vice versa. These analyses are conducted for the following tasks with sufficient language pairs: BLI with PanLex, parsing, and POS tagging.
For these analyses we report average correlation: across target languages in the source-language selection analysis, or across the source languages in the target-language selection analysis. We provide the percentage of times each compared measure scored the highest for a particular task and setting.
Stepwise regression analysis is not suitable for the selection analysis due to the limited number of language pairs in each language selection setup (e.g., PanLex offers 14 language pairs for each source- or target- language selection analysis). These conditions impede the statistical significance power of the tests which stepwise regression requires. We therefore opt for a standard multiple linear regression model instead; the regressors include the isomorphism measure with the highest individual correlation combined with the linguistic measures. Similarly to the stepwise analysis, we report the unified correlation coefficient, .
We observe that the same findings reported for the aggregated correlation analyses (Tables 1 and 2 in the main text) also hold for the language selection analyses (Table 3 below). This indicates that our spectral measures have an applicative value as they can facilitate better cross-language transfer.
We also observe interesting patterns in the selection analyses for the POS tagging task in Table 3: While the results in the target-language selection analysis largely follow the main-text results, the same does not hold for source-language selection (Table 3, POS Target and Source columns). We speculate that this is in fact an artefact of the experimental design of Lin et al. (2019). Their set of target languages deliberately comprises only truly low-resource languages, and such languages are expected to have lower-quality embedding spaces. Transferring to such languages is bound to fail with most source languages regardless of the actual source-target language similarity. The difficulty of this setting is reflected in the actual scores: average accuracy scores for the best source-target combination is 0.55 in the source-language selection analysis, and 0.92 for target-language selection.
Appendix C Single and Combined Analysis
We show (Figure 1) a single experimental condition (Sup method in the PanLex BLI dataset, leftmost column in Table 1 in the main text) to illustrate the data distribution and the correlation for spectral-based measures (e.g., econd-hm), and their improvement once this measure is combined with linguistic measures through regression analysis.
Task | BLI | DEP | POS | |||||||
Selection | Target | Source | Target | Source | Target | Source | ||||
Sup | Sup+ | Unsup | Sup | Sup+ | Unsup | |||||
svg | -.5731% | -.5631% | -.5435% | -.5328% | -.5429% | -.5638% | -.6256% | -.7368% | -.514% | -.10 |
cond-m | -.59 | -.55 | -.42 | -.51 | -.49 | -.38 | -.44 | -.64 | -.58 | -.08 |
econd-m | -.6431% | -.5725% | -.4417% | -.5425% | -.5324% | -.4317% | -.513% | -.706% | -.6110% | .05 |
gh | -.32 | -.333% | -.277% | -.31 | -.28 | -.26 | -.5116% | -.6313% | -.5834% | -.11 |
is | -.30 | -.30 | -.303% | -.333% | -.34 | -.327% | -.5216% | -.5810% | -.4350% | -.23 |
phy | -.257% | -.273% | -.233% | -.4124% | -.4234% | -.3210% | -.44 | -.42 | -.14 | -.14 |
typ | -.5024% | -.4928% | -.4228% | -.4517% | -.4110% | -.3621% | -.403% | -.42 | -.182% | -.05 |
geo | -.297% | -.3410% | -.327% | -.313% | -.323% | -.327% | -.486% | -.563% | -.12 | -.33 |
.86 | .83 | .76 | .83 | .83 | .77 | .79 | .85 | .68 | – |


Language family | Language | PanLex | MUSE | GTrans | MT | DEP | POS |
---|---|---|---|---|---|---|---|
Germanic (IE) | English | 28 | 88 | 7 | 54 | 62 | 17 |
Germanic (IE) | German | 28 | 10 | 7 | 1 | 62 | 17 |
Germanic (IE) | Dutch | 28 | 2 | 1 | 62 | 17 | |
Germanic (IE) | Swedish | 28 | 2 | 1 | 62 | 17 | |
Germanic (IE) | Danish | 28 | 2 | 1 | 62 | 17 | |
Germanic (IE) | Norwegian | 28 | 2 | 1 | 62 | 17 | |
Germanic (IE) | Afrikaans | 2 | 65 | ||||
Germanic (IE) | Faroese | 50 | |||||
Romance (IE) | Italian | 28 | 10 | 7 | 1 | 62 | 17 |
Romance (IE) | Portuguese | 28 | 10 | 1 | 62 | 17 | |
Romance (IE) | Spanish | 28 | 10 | 1 | 62 | 17 | |
Romance (IE) | French | 28 | 10 | 7 | 1 | 62 | 17 |
Romance (IE) | Romanian | 28 | 2 | 1 | 62 | 17 | |
Romance (IE) | Catalan | 28 | 2 | 62 | 17 | ||
Romance (IE) | Galician | 1 | 17 | ||||
Romance (IE) | Latin | 62 | 17 | ||||
Slavic (IE) | Croatian | 28 | 2 | 7 | 1 | 62 | 17 |
Slavic (IE) | Polish | 28 | 2 | 1 | 62 | 17 | |
Slavic (IE) | Russian | 28 | 2 | 7 | 1 | 62 | 17 |
Slavic (IE) | Czech | 28 | 2 | 1 | 62 | 17 | |
Slavic (IE) | Bulgarian | 56 | 2 | 1 | 62 | 17 | |
Slavic (IE) | Bosnian | 2 | 1 | ||||
Slavic (IE) | Macedonian | 2 | 1 | ||||
Slavic (IE) | Slovak | 2 | 1 | 62 | 17 | ||
Slavic (IE) | Slovenian | 2 | 1 | 62 | 17 | ||
Slavic (IE) | Ukrainian | 2 | 1 | 62 | 17 | ||
Slavic (IE) | Belarusian | 1 | 65 | ||||
Slavic (IE) | Serbian | 1 | 17 | ||||
Hellenic (IE) | Modern Greek | 2 | 1 | 17 | |||
Balto-Slavic (IE) | Lithuanian | 28 | 2 | 1 | 65 | ||
Balto-Slavic (IE) | Latvian | 2 | 62 | 17 | |||
Indo-Iranian (IE) | Bengali | 2 | 1 | ||||
Indo-Iranian (IE) | Persian | 2 | 1 | 17 | |||
Indo-Iranian (IE) | Hindi | 2 | 1 | 62 | 17 | ||
Indo-Iranian (IE) | Kurdish | 1 | |||||
Indo-Iranian (IE) | Marathi | 1 | 65 | ||||
Indo-Iranian (IE) | Urdu | 1 | 17 | ||||
Indo-Iranian (IE) | Sanskrit | 50 | |||||
Celtic (IE) | Irish | 65 | |||||
Celtic (IE) | Breton | 50 | |||||
Albanian (IE) | Albanian | 2 | 1 | ||||
Armenic (IE) | Armenian | 1 | 65 | ||||
Turkic | Turkish | 28 | 2 | 7 | 1 | 17 | |
Turkic | Azerbaijani | 1 | |||||
Turkic | Kazakh | 1 | 65 | ||||
Turkic | Uighur | 17 | |||||
Semitic | Hebrew | 28 | 2 | 1 | 62 | 17 | |
Semitic | Arabic | 2 | 1 | 62 | 17 | ||
Semitic | Amharic | 50 | |||||
Uralic | Hungarian | 28 | 2 | 1 | 65 | ||
Uralic | Finnish | 28 | 2 | 7 | 1 | 62 | 17 |
Uralic | Estonian | 28 | 2 | 1 | 62 | 17 | |
Austronesian | Indonesian | 28 | 2 | 1 | 62 | 17 | |
Austronesian | Malay | 2 | 1 | ||||
Austronesian | Tagalog | 2 | 50 | ||||
Dravidian | Tamil | 2 | 1 | 65 | |||
Dravidian | Telugu | 65 | |||||
Sino-Tibetan | Chinese | 2 | 1 | 62 | 17 | ||
Sino-Tibetan | Burmese | 1 | |||||
Japonic | Japanese | 2 | 1 | 62 | 17 | ||
Kartvelian | Georgian | 28 | 1 | ||||
Koreanic | Korean | 28 | 2 | 1 | 62 | 17 | |
Mongolic | Mongolian | 1 | |||||
Niger-Congo | Yoruba | 50 | |||||
Austroasiatic | Vietnamese | 2 | 1 | 17 | |||
Tai-Kadai | Thai | 28 | 2 | 1 | 50 | ||
Isolate | Basque | 28 | 1 | 17 | |||
Not defined | Esperanto | 28 | 1 |