-
Notifications
You must be signed in to change notification settings - Fork 284
Translating Expert Radiology Reports into Layman Summaries using pre-trained Flan-T5 Model. Task #13 #3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: topic-recognition
Are you sure you want to change the base?
Conversation
…for argument purposes.
…gFace without their API. Still under testing.
…values for Seq2Seq training. Optimiser used is AdamW, as used before in demos.
…s is sampled from total "train" dataset
…for training and validation respectively
… training loop from pytorch.org and huggingface tutorials.
… most basic implementation. Rouge scores are printed at the end of each epoch.
…nt of rows (not random, starting from 0)
…fficial layman report in the dataset to fine tuned model.
…uses the HuggingFace Dataset API for convenience (might change to custom dataset later)
Implemented all sections except for benchmark examples. Minor improvements and grammatical checks required.
…for real training.
Set default benchmark to 5 samples from validation split.
…ly or via rangpur.
Added note about training completion and hardware requirements.
Corrected 'remove' to 'remote' in training note.
Added constants for dataset links, training parameters, and model prompt.
Added docstrings to methods for better documentation.
Added docstrings to FlanModel methods for clarity.
Added a comment to clarify the perplexity calculation.
Added detailed docstrings for training and evaluation methods, improved comments for clarity, and included optimizer setup in the training script.
|
Recognition Problem : total : 20 |
|
s4885380 |
Marking
Marked as per the due date and changes after which aren't necessarily allowed to contribute to grade for fairness. |
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Expert to Layman Radiology Reports
Overview:
This project aims to fine-tune an existing encoder-decoder LLM - the Flan-T5 in specifically summarising data from the
BioLaySumm 2025dataset.Fine-tuning was performed on the
rangpurcluster, with full training on the NVDIA A100 GPU.Files:
constants.py- File containing all parameters and other constant variablesdataset.py- Custom dataset loader usingPandasfor model training and evaluationmodules.py- Importer for the pre-trainedFlan-T5modelpredict.py- Model benchmarking on unseenvalidationdata splitrequirements.txt- List of all dependencies and their respective versiontrain.py- Custom training and evaluation loop on thetraindata splitassets/- Folder for allREADMErequired imagesrunners/- Folder containingtrainandbenchmarkscripts for use onrangpurResult of Model Fine-tuning:
rougescores:rouge1: 71.19rouge2: 51.83rougeL: 65.80rougeLsum: 65.82perplexityscores (extra metrics):perplexityscore is consistently< 3