a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
-
Updated
Jun 25, 2024 - Jupyter Notebook
a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
This repository features a custom-built decoder-only language model (LLM) with a total of 37 million parameters 🔥. I train the model to be able to ask question from a given context
Experimental project for AI and NLP based on Transformer Architecture
Generate caption on images using CNN Encoder- LSTM Decoder structure
Transformers Intuition
Implementation of the GPT-3 paper: Language Models are Few-Shot Learners
Generative AI fine-tune and inference for sequence classification tasks
An explainable and simplified version of OLMo model
a dna sequence generation/classification using transformers
An LLM based tool for generation of cheese advirtisements
This project aims to simplify texts from research papers using advanced natural language processing (NLP) techniques, making them more accessible to a broader audience
Coding A Decoder Only Transformer Like ChatGPT From Scratch
Using LLMs in huggingface for sentiment analysis, translation, summarization and extractive question answering
On the Design and Performance of Machine Learning Based Error Correcting Decoders
Custom decoder Transformer that treats a patient's medical journey like a story told through diagnosis codes instead of words.
Decoder model for language modelling
Build Text summarizer for arabic language
A miniGPT inspired from the original NanoGPT released by OpenAI. This is a notebook to walk through the decoder part of the transformer architecture with details outlined.
Text Generation using RNN, LSTM, and Transformer
Add a description, image, and links to the decoder-model topic page so that developers can more easily learn about it.
To associate your repository with the decoder-model topic, visit your repo's landing page and select "manage topics."