#自然语言处理#Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
翻译 - PyTorch 中的开源预训练模型框架和预训练模型 Zoo
#自然语言处理#Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
#自然语言处理#🤖 A PyTorch library of curated Transformer models and their composable components
#自然语言处理#Trankit is a Light-Weight Transformer-based Python Toolkit for Multilingual Natural Language Processing
翻译 - Trankit是用于多语言自然语言处理的基于轻型变压器的Python工具包
This repository contains the official release of the model "BanglaBERT" and associated downstream finetuning code and datasets introduced in the paper titled "BanglaBERT: Language Model Pretraining an...
#自然语言处理#CINO: Pre-trained Language Models for Chinese Minority (少数民族语言预训练模型)
#大语言模型#Unattended Lightweight Text Classifiers with LLM Embeddings
Neural syntax annotator, supporting sequence labeling, lemmatization, and dependency parsing.
#自然语言处理#Deep-learning system proposed by HFL for SemEval-2022 Task 8: Multilingual News Similarity
#自然语言处理#Resources and tools for the Tutorial - "Hate speech detection, mitigation and beyond" presented at ICWSM 2021
PyTorch implementation of Sentiment Analysis of the long texts written in Serbian language (which is underused language) using pretrained Multilingual RoBERTa based model (XLM-R) on the small dataset.
Sentiment Analysis of tweets written in underused Slavic languages (Serbian, Bosnian and Croatian) using pretrained multilingual RoBERTa based model XLM-R on 2 different datasets.
An implementation of drophead regularization for pytorch transformers
This is a Pytorch (+ Huggingface transformers) implementation of a "simple" text classifier defined using BERT-based models. In this lab we will see how it is simple to use BERT for a sentence classif...
Improving Low-Resource Neural Machine Translation of Related Languages by Transfer Learning
#自然语言处理#This repository is a comprehensive project that leverages the XLM-Roberta model for intent detection. This repository is a valuable resource for developers looking to build and fine-tune intent detect...
Improving Bilingual Lexicon Induction with Cross-Encoder Reranking (Findings of EMNLP 2022). Keywords: Bilingual Lexicon Induction, Word Translation, Cross-Lingual Word Embeddings.
notebooks to finetune `bert-small-amharic`, `bert-mini-amharic`, and `xlm-roberta-base` models using an Amharic text classification dataset and the transformers library
#自然语言处理#Language Model Decomposition: Quantifying the Dependency and Correlation of Language Models
#自然语言处理#A case study of NLI ( Natural Language Inferencing) with Transfer Learning. Kaggle Competition Rank - 18th (Global)