Skip to content

Conversation

@piradiusquared
Copy link

@piradiusquared piradiusquared commented Oct 14, 2025

Expert to Layman Radiology Reports

Overview:

This project aims to fine-tune an existing encoder-decoder LLM - the Flan-T5 in specifically summarising data from the BioLaySumm 2025 dataset.
Fine-tuning was performed on the rangpur cluster, with full training on the NVDIA A100 GPU.

Files:

constants.py - File containing all parameters and other constant variables
dataset.py - Custom dataset loader using Pandas for model training and evaluation
modules.py - Importer for the pre-trained Flan-T5 model
predict.py - Model benchmarking on unseen validation data split
requirements.txt - List of all dependencies and their respective version
train.py - Custom training and evaluation loop on the train data split
assets/ - Folder for all README required images
runners/ - Folder containing train and benchmark scripts for use on rangpur

Result of Model Fine-tuning:

rouge scores:

  • rouge1: 71.19
  • rouge2: 51.83
  • rougeL: 65.80
  • rougeLsum: 65.82

perplexity scores (extra metrics):

  • On average, the fine-tuned model is 2 - 4 times more confident in selecting the next appropriate token (word) for summaries.
  • Average perplexity score is consistently < 3

@piradiusquared piradiusquared changed the title Initial Pull Request FLAN-T5 Laymann Oct 16, 2025
@piradiusquared piradiusquared changed the title FLAN-T5 Laymann FLAN-T5 Layman Oct 16, 2025
…gFace without their API. Still under testing.
…values for Seq2Seq training.

Optimiser used is AdamW, as used before in demos.
… training loop from pytorch.org and huggingface tutorials.
… most basic implementation. Rouge scores are printed at the end of each epoch.
…fficial layman report in the dataset to fine tuned model.
@piradiusquared piradiusquared changed the title FLAN-T5 Layman Translating Expert Radiology Reports into Layman Summaries using pre-trained Flan-T5 Model. Task #13 Oct 31, 2025
Added constants for dataset links, training parameters, and model prompt.
Added docstrings to methods for better documentation.
Added docstrings to FlanModel methods for clarity.
Added a comment to clarify the perplexity calculation.
Added detailed docstrings for training and evaluation methods, improved comments for clarity, and included optimizer setup in the training script.
@Claire1217
Copy link

Recognition Problem : total : 20
Solves problem: Good data augmentation design. Very good reasoning from test run 1 to test run 2. (5)
Implementation functions : Well-organized codes. Good practice to have your constants in an individual file(3)
Good design: good (1)
Commenting: Well commented (1)
Difficulty: Hard (10)

@wangzhaomxy
Copy link
Collaborator

s4885380

@gayanku
Copy link
Collaborator

gayanku commented Nov 24, 2025

Marking

Good/OK/Fair Practice (Design/Commenting, TF/Torch Usage)
Good design and implementation.
Spacing and comments.
No Header blocks. -1
Recognition Problem
Good solution to problem.
Driver Script present.
File structure present.
Good Usage & Demo & Visualisation & Data usage.
Module present.
Commenting present.
No Data leakage found.
Difficulty : Hard. Hard Difficulty : LLM
Commit Log
Good Meaningful commit messages.
Good Progressive commits.
Documentation
Readme :Good.
Model/technical explanation :Good.
Description and Comments :Good.
Markdown used and PDF submitted.
Pull Request
Successful Pull Request (Working Algorithm Delivered on Time in Correct Branch).
No Feedback required.
Request Description is good.
TOTAL-1

Marked as per the due date and changes after which aren't necessarily allowed to contribute to grade for fairness.
Subject to approval from Shakes

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants