Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
-
Updated
Nov 10, 2021 - Python
Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
Match celebrity users with their respective tweets by making use of Semantic Textual Similarity on over 900+ celebrity users' 2.5 million+ scraped tweets utilizing SBERT, streamlit, tweepy and FastAPI
Language Model Decomposition: Quantifying the Dependency and Correlation of Language Models
Code and data for the EMNLP 2020 paper: "Detecting Fine-Grained Cross-Lingual Semantic Divergences without Supervision by Learning to Rank"
In this project, we compared Spanish BERT and Multilingual BERT in the Sentiment Analysis task.
Zero-shot and Translation Experiments on XQuAD, MLQA and TyDiQA
This repository contains a number of experiments with Multi Lingual Transformer models (Multi-Lingual BERT, DistilBERT, XLM-RoBERTa, mT5 and ByT5) focussed on the Dutch language.
This work was done as part of Kaggle InClass competition hosted by Sharechat. We had to develop AI solutions for predicting abusive comments posted on the Moj app in 10+ languages given natural language data and user context data.
Cross-lingual misinformation detection
By using the hypothesis of historical linguistics, we found a way to improve the performance of multilingual transformers with limited amount of data
Align Parallel Sentence of 104 Languages with the help of mBERT and LaBSE
This repository implements a Multilingual BERT (mBERT) model for performing Parts-of-Speech (POS) Tagging on Assamese-English code-mixed texts.
submission from team import-winning-model
CS224n Homeworks and final project
Add a description, image, and links to the multilingual-bert topic page so that developers can more easily learn about it.
To associate your repository with the multilingual-bert topic, visit your repo's landing page and select "manage topics."