[NAACL 2021] This is the code for our paper `Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach'.
-
Updated
Aug 17, 2022 - Python
[NAACL 2021] This is the code for our paper `Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach'.
The purpose of this repository is to introduce new dialogue-level commonsense inference datasets and tasks. We chose dialogues as the data source because dialogues are known to be complex and rich in commonsense.
Topic clustering library built on Transformer embeddings and cosine similarity metrics.Compatible with all BERT base transformers from huggingface.
This app Classifies the text generated by AI tools like chatGPT. Roberta-base-openai-detector Model has been used from hugging face to detect ai generated texts.
This repository contains the code of our winning solution for the Shared Task on Detecting Signs of Depression from Social Media Text at LT-EDI-ACL2022.
Source code for CoNLL 2021 paper by Huebner et al. 2021
Convert pretrained RoBerta models to various long-document transformer models
This repository contains the solutions to three problem statements completed during the hackathon. Each problem statement is categorized based on its difficulty level: Easy, Moderate, and Hard.
It is the nlp task to classify empathetic dialogues datasets using RoBERTa, ERNIE-2.0 and XLNet with different preprocessing method. You can get some detailed introduction and experimental results in the link below.
Building this project to generate MCQ Questions from any type of text and generate answers and distractors for it.
Resources for the paper: Monolingual Pre-trained Language Models for Tigrinya
Tutorial on training a RoBERTa Transformers model from scratch
🙂🙃 Being happy :) being sad :( with this tool, you become sentiment GIGA chad!
Generating code with Ludwig AI/ML (PyTorch, Tensorflow)
This work focuses on the development of machine learning models, in particular neural networks and SVM, where they can detect toxicity in comments. The topics we will be dealing with: a) Cost-sensitive learning, b) Class imbalance
predict movie's genres based on overview
This repository contains the code, models and corpus of the project "Generative Adversarial Networks for Text-to-Image Synthesis & Generation: A Comparative Analysis of Natural Language Processing models for the Spanish language".
Studies show that people are more depressed than ever after the pandemic, but is the way we are measuring depression even accurate?
A project demonstrating the use of Large Language Models (LLMs) for text classification using the RoBERTa model.
Add a description, image, and links to the roberta-model topic page so that developers can more easily learn about it.
To associate your repository with the roberta-model topic, visit your repo's landing page and select "manage topics."