Course materials for the Machine Learning for NLP course taught by Sameer Singh for the Cognitive Science summer school 2022.
-
Updated
Aug 7, 2022 - Jupyter Notebook
Course materials for the Machine Learning for NLP course taught by Sameer Singh for the Cognitive Science summer school 2022.
Fine-tuning (or training from scratch) the library models for language modeling on a text dataset for GPT, GPT-2, ALBERT, BERT, DitilBERT, RoBERTa, XLNet... GPT and GPT-2 are trained or fine-tuned using a causal language modeling (CLM) loss while ALBERT, BERT, DistilBERT and RoBERTa are trained or fine-tuned using a masked language modeling (MLM…
An AI generated picturebook.
This is the implementation of low rank adaptation (LoRA) which is a subset of parameter efficient fine tuning (PEFT).
Repository for My HuggingFace Natural Language Processing Projects
Transformers Intuition
Causal language modeling and intent classification using GPT-2.
Dataset and model fine-tuning for function calling
Rescoring Automatic Speech Recognition using Large Language Models
A quick and easy way to interact with open-source LLMs.
Auto-regressive causal language model for molecule (SMILES) and reaction template (SMARTS) generation based on the Hugging Face implementation of OpenAI's GPT-2 transformer decoder model
Links to my repositories, where I implement a wide variety of Natural Language Processing models using TensorFlow and Hugging Face.
This repository is for the paper Lexical Substitution as Causal Language Modeling. In Proceedings of the 13th Joint Conference on Lexical and Computational Semantics (*SEM 2024), Mexico City, Mexico. Association for Computational Linguistics.
Codebase for arXiv:2405.17767, based on GPT-Neo and TinyStories.
Add a description, image, and links to the causal-language-modeling topic page so that developers can more easily learn about it.
To associate your repository with the causal-language-modeling topic, visit your repo's landing page and select "manage topics."