Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
-
Updated
Nov 10, 2021 - Python
Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
Language Model Decomposition: Quantifying the Dependency and Correlation of Language Models
Code and data for the EMNLP 2020 paper: "Detecting Fine-Grained Cross-Lingual Semantic Divergences without Supervision by Learning to Rank"
Codes for the short essay project for Mphil TAL by Yingjia Wan in 2023.
This repository contains a number of experiments with Multi Lingual Transformer models (Multi-Lingual BERT, DistilBERT, XLM-RoBERTa, mT5 and ByT5) focussed on the Dutch language.
Match celebrity users with their respective tweets by making use of Semantic Textual Similarity on over 900+ celebrity users' 2.5 million+ scraped tweets utilizing SBERT, streamlit, tweepy and FastAPI
Align Parallel Sentence of 104 Languages with the help of mBERT and LaBSE
Add a description, image, and links to the multilingual-bert topic page so that developers can more easily learn about it.
To associate your repository with the multilingual-bert topic, visit your repo's landing page and select "manage topics."