Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
-
Updated
Nov 10, 2021 - Python
Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
Match celebrity users with their respective tweets by making use of Semantic Textual Similarity on over 900+ celebrity users' 2.5 million+ crawled tweets utilizing SBERT, streamlit, tweepy and FastAPI
Language Model Decomposition: Quantifying the Dependency and Correlation of Language Models
Code and data for the EMNLP 2020 paper: "Detecting Fine-Grained Cross-Lingual Semantic Divergences without Supervision by Learning to Rank"
This repository contains a number of experiments with Multi Lingual Transformer models (Multi-Lingual BERT, DistilBERT, XLM-RoBERTa, mT5 and ByT5) focussed on the Dutch language.
Bias analysis in multilingual BERT using SHAP and fairness metrics (EN, DE, HI)
App for medieval multilingual metadata — feeds the Segmentation Dataset and trains the Aquilign aligner
Align Parallel Sentence of 104 Languages with the help of mBERT and LaBSE
This repository implements a Multilingual BERT (mBERT) model for performing Parts-of-Speech (POS) Tagging on Assamese-English code-mixed texts.
Add a description, image, and links to the multilingual-bert topic page so that developers can more easily learn about it.
To associate your repository with the multilingual-bert topic, visit your repo's landing page and select "manage topics."