Hello. I have been looking around to figure out the usage of pretrained embeddings for multiple languages. What I am trying to achieve is to have pretrained embeddings for multiple languages so that my ML model knows similar words across multiple languages. For example, "Good" in English is the same as "Gut" in German. Any help in this regard would be highly appreciated
I think you should check the relevant papers from Mikel Artetxe:
And their implementation in the vecmap library :)
And the MUSE library contains several pretrained word embeddings for English-X :)
@stefan-it Thank you very much for replying so quickly. I will check the papers and vecmap library as you pointed out. I have a question about MUSE though. For instance, German-English there is an entry "mit with", would that mean that I can use the embeddings of with for mit, or vice versa, so that the similar words can be clustered?
I haven't tried it yet, but there's a nice Notebook that shows how to get nearest neighbors and even visualize bilingual embeddings:

See here:
https://github.com/facebookresearch/MUSE/blob/master/demo.ipynb
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.