Implementation and demo of explainable coding of clinical notes with Hierarchical Label-wise Attention Networks (HLAN)
-
Updated
Jul 15, 2022 - Python
Implementation and demo of explainable coding of clinical notes with Hierarchical Label-wise Attention Networks (HLAN)
This repository contains code for a fine-tuning experiment of CamemBERT, a French version of the BERT language model, on a portion of the FQuAD (French Question Answering Dataset) for Question Answering tasks.
Very Simple Transformers provides a simplified interface for packaging, deploying, and serving Transformer models.
Application for training the pretrained transformer model DeBERTaV3 on an Aspect Based Sentiment Analysis task
Deep learning in FEN’s win / loss evaluation.
Deep learning in smiles win / loss evaluation.
This library is based on simpletransformers and HuggingFace's Transformers library.
Weak Supervised Fake News Detection with RoBERTa, XLNet, ALBERT, XGBoost and Logistic Regression classifiers.
Small application to test out some functionality of OpenAIs Generative Pre-Trained Transformer (GPT-2) Model
Simple Transformers Fork that supports T5TokenizerFast and umT5
Backend for MindPeers ML (NLP) models such as Sentiment Analysis & Keyword Extraction (including Feedback Loops)
Add a description, image, and links to the simpletransformers topic page so that developers can more easily learn about it.
To associate your repository with the simpletransformers topic, visit your repo's landing page and select "manage topics."