Skip to content
This repository has been archived by the owner on Dec 16, 2022. It is now read-only.

FairScale integration and T5-11B fine-tuning #271

Merged
merged 29 commits into from
Jul 19, 2021
Merged

FairScale integration and T5-11B fine-tuning #271

merged 29 commits into from
Jul 19, 2021

Conversation

epwalsh
Copy link
Member

@epwalsh epwalsh commented Jun 3, 2021

Corresponding PR for allenai/allennlp#5242

@epwalsh epwalsh mentioned this pull request Jun 3, 2021
11 tasks
Makefile Outdated Show resolved Hide resolved
requirements.txt Outdated Show resolved Hide resolved
@epwalsh epwalsh changed the title [WIP] FairScale integration and T5-11B fine-tuning FairScale integration and T5-11B fine-tuning Jun 29, 2021
@epwalsh epwalsh marked this pull request as ready for review June 29, 2021 22:25
@epwalsh epwalsh requested review from dirkgr and AkshitaB June 29, 2021 22:25
Comment on lines 20 to 21
// AMP is currently unusably slow with t5-11b, which be due to a bug bug within
// FairScale, but I'm not sure yet.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Typo

Comment on lines 65 to 73
local wandb_callback = {
"type": "wandb",
"project": "allennlp-t5",
"entity": "allenai-team1",
"watch_model": false,
"summary_interval": 1,
"should_log_parameter_statistics": false,
"should_log_learning_rate": false,
};
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I don't know if we want this in a default config. Virtually everyone using this will have to change it.

Though I would be entirely in favor of having a commented-out wandb config in every jsonnet file.

[if !debug then "callbacks"]: [wandb_callback],
},
"distributed": {
"cuda_devices": if debug then [0, 1] else [0, 1, 2, 3, 4, 5, 6, 7],
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@@ -0,0 +1,125 @@
// =================== Configurable Settings ======================
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The filename is confusing if this works for both 11b models and smaller ones.

@epwalsh epwalsh merged commit db0e21a into main Jul 19, 2021
@epwalsh epwalsh deleted the fairscale branch July 19, 2021 23:39
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
2 participants