-
Notifications
You must be signed in to change notification settings - Fork 181
/
run_train.py
781 lines (734 loc) · 31.2 KB
/
run_train.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
###########################################################################################
# Training script for MACE
# Authors: Ilyes Batatia, Gregor Simm, David Kovacs
# This program is distributed under the MIT License (see MIT.md)
###########################################################################################
import argparse
import ast
import glob
import json
import logging
import os
from copy import deepcopy
from pathlib import Path
from typing import List, Optional
import torch.distributed
import torch.nn.functional
from e3nn.util import jit
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.utils.data import ConcatDataset
from torch_ema import ExponentialMovingAverage
import mace
from mace import data, tools
from mace.calculators.foundations_models import mace_mp, mace_off
from mace.tools import torch_geometric
from mace.tools.model_script_utils import configure_model
from mace.tools.multihead_tools import (
HeadConfig,
assemble_mp_data,
dict_head_to_dataclass,
prepare_default_head,
)
from mace.tools.scripts_utils import (
LRScheduler,
convert_to_json_format,
create_error_table,
dict_to_array,
extract_config_mace_model,
get_atomic_energies,
get_avg_num_neighbors,
get_config_type_weights,
get_dataset_from_xyz,
get_files_with_suffix,
get_loss_fn,
get_optimizer,
get_params_options,
get_swa,
print_git_commit,
setup_wandb,
)
from mace.tools.slurm_distributed import DistributedEnvironment
from mace.tools.utils import AtomicNumberTable
def main() -> None:
"""
This script runs the training/fine tuning for mace
"""
args = tools.build_default_arg_parser().parse_args()
run(args)
def run(args: argparse.Namespace) -> None:
"""
This script runs the training/fine tuning for mace
"""
tag = tools.get_tag(name=args.name, seed=args.seed)
args, input_log_messages = tools.check_args(args)
if args.device == "xpu":
try:
import intel_extension_for_pytorch as ipex
except ImportError as e:
raise ImportError(
"Error: Intel extension for PyTorch not found, but XPU device was specified"
) from e
if args.distributed:
try:
distr_env = DistributedEnvironment()
except Exception as e: # pylint: disable=W0703
logging.error(f"Failed to initialize distributed environment: {e}")
return
world_size = distr_env.world_size
local_rank = distr_env.local_rank
rank = distr_env.rank
if rank == 0:
print(distr_env)
torch.distributed.init_process_group(backend="nccl")
else:
rank = int(0)
# Setup
tools.set_seeds(args.seed)
tools.setup_logger(level=args.log_level, tag=tag, directory=args.log_dir, rank=rank)
logging.info("===========VERIFYING SETTINGS===========")
for message, loglevel in input_log_messages:
logging.log(level=loglevel, msg=message)
if args.distributed:
torch.cuda.set_device(local_rank)
logging.info(f"Process group initialized: {torch.distributed.is_initialized()}")
logging.info(f"Processes: {world_size}")
try:
logging.info(f"MACE version: {mace.__version__}")
except AttributeError:
logging.info("Cannot find MACE version, please install MACE via pip")
logging.debug(f"Configuration: {args}")
tools.set_default_dtype(args.default_dtype)
device = tools.init_device(args.device)
commit = print_git_commit()
model_foundation: Optional[torch.nn.Module] = None
if args.foundation_model is not None:
if args.multiheads_finetuning:
assert (
args.E0s != "average"
), "average atomic energies cannot be used for multiheads finetuning"
if args.foundation_model in ["small", "medium", "large"]:
logging.info(
f"Using foundation model mace-mp-0 {args.foundation_model} as initial checkpoint."
)
calc = mace_mp(
model=args.foundation_model,
device=args.device,
default_dtype=args.default_dtype,
)
model_foundation = calc.models[0]
elif args.foundation_model in ["small_off", "medium_off", "large_off"]:
model_type = args.foundation_model.split("_")[0]
logging.info(
f"Using foundation model mace-off-2023 {model_type} as initial checkpoint. ASL license."
)
calc = mace_off(
model=model_type,
device=args.device,
default_dtype=args.default_dtype,
)
model_foundation = calc.models[0]
else:
model_foundation = torch.load(
args.foundation_model, map_location=args.device
)
logging.info(
f"Using foundation model {args.foundation_model} as initial checkpoint."
)
args.r_max = model_foundation.r_max.item()
else:
args.multiheads_finetuning = False
if args.heads is not None:
args.heads = ast.literal_eval(args.heads)
else:
args.heads = prepare_default_head(args)
logging.info("===========LOADING INPUT DATA===========")
heads = list(args.heads.keys())
logging.info(f"Using heads: {heads}")
head_configs: List[HeadConfig] = []
for head, head_args in args.heads.items():
logging.info(f"============= Processing head {head} ===========")
head_config = dict_head_to_dataclass(head_args, head, args)
if head_config.statistics_file is not None:
with open(head_config.statistics_file, "r") as f: # pylint: disable=W1514
statistics = json.load(f)
logging.info("Using statistics json file")
head_config.r_max = (
statistics["r_max"] if args.foundation_model is None else args.r_max
)
head_config.atomic_numbers = statistics["atomic_numbers"]
head_config.mean = statistics["mean"]
head_config.std = statistics["std"]
head_config.avg_num_neighbors = statistics["avg_num_neighbors"]
head_config.compute_avg_num_neighbors = False
if isinstance(statistics["atomic_energies"], str) and statistics[
"atomic_energies"
].endswith(".json"):
with open(statistics["atomic_energies"], "r", encoding="utf-8") as f:
atomic_energies = json.load(f)
head_config.E0s = atomic_energies
head_config.atomic_energies_dict = ast.literal_eval(atomic_energies)
else:
head_config.E0s = statistics["atomic_energies"]
head_config.atomic_energies_dict = ast.literal_eval(
statistics["atomic_energies"]
)
# Data preparation
if head_config.train_file.endswith(".xyz"):
if head_config.valid_file is not None:
assert head_config.valid_file.endswith(
".xyz"
), "valid_file if given must be same format as train_file"
config_type_weights = get_config_type_weights(
head_config.config_type_weights
)
collections, atomic_energies_dict = get_dataset_from_xyz(
work_dir=args.work_dir,
train_path=head_config.train_file,
valid_path=head_config.valid_file,
valid_fraction=head_config.valid_fraction,
config_type_weights=config_type_weights,
test_path=head_config.test_file,
seed=args.seed,
energy_key=head_config.energy_key,
forces_key=head_config.forces_key,
stress_key=head_config.stress_key,
virials_key=head_config.virials_key,
dipole_key=head_config.dipole_key,
charges_key=head_config.charges_key,
head_name=head_config.head_name,
keep_isolated_atoms=head_config.keep_isolated_atoms,
)
head_config.collections = collections
head_config.atomic_energies_dict = atomic_energies_dict
logging.info(
f"Total number of configurations: train={len(collections.train)}, valid={len(collections.valid)}, "
f"tests=[{', '.join([name + ': ' + str(len(test_configs)) for name, test_configs in collections.tests])}],"
)
head_configs.append(head_config)
if all(head_config.train_file.endswith(".xyz") for head_config in head_configs):
size_collections_train = sum(
len(head_config.collections.train) for head_config in head_configs
)
size_collections_valid = sum(
len(head_config.collections.valid) for head_config in head_configs
)
if size_collections_train < args.batch_size:
logging.error(
f"Batch size ({args.batch_size}) is larger than the number of training data ({size_collections_train})"
)
if size_collections_valid < args.valid_batch_size:
logging.warning(
f"Validation batch size ({args.valid_batch_size}) is larger than the number of validation data ({size_collections_valid})"
)
if args.multiheads_finetuning:
logging.info(
"==================Using multiheads finetuning mode=================="
)
args.loss = "universal"
if (
args.foundation_model in ["small", "medium", "large"]
or "mp" in args.foundation_model
or args.pt_train_file is None
):
logging.info(
"Using foundation model for multiheads finetuning with Materials Project data"
)
heads = list(dict.fromkeys(["pt_head"] + heads))
head_config_pt = HeadConfig(
head_name="pt_head",
E0s="foundation",
statistics_file=args.statistics_file,
compute_avg_num_neighbors=False,
avg_num_neighbors=model_foundation.interactions[0].avg_num_neighbors,
)
collections = assemble_mp_data(args, tag, head_configs)
head_config_pt.collections = collections
head_config_pt.train_file = f"mp_finetuning-{tag}.xyz"
head_configs.append(head_config_pt)
else:
logging.info(
f"Using foundation model for multiheads finetuning with {args.pt_train_file}"
)
heads = list(dict.fromkeys(["pt_head"] + heads))
collections, atomic_energies_dict = get_dataset_from_xyz(
work_dir=args.work_dir,
train_path=args.pt_train_file,
valid_path=args.pt_valid_file,
valid_fraction=args.valid_fraction,
config_type_weights=None,
test_path=None,
seed=args.seed,
energy_key=args.energy_key,
forces_key=args.forces_key,
stress_key=args.stress_key,
virials_key=args.virials_key,
dipole_key=args.dipole_key,
charges_key=args.charges_key,
head_name="pt_head",
keep_isolated_atoms=args.keep_isolated_atoms,
)
head_config_pt = HeadConfig(
head_name="pt_head",
train_file=args.pt_train_file,
valid_file=args.pt_valid_file,
E0s="foundation",
statistics_file=args.statistics_file,
valid_fraction=args.valid_fraction,
config_type_weights=None,
energy_key=args.energy_key,
forces_key=args.forces_key,
stress_key=args.stress_key,
virials_key=args.virials_key,
dipole_key=args.dipole_key,
charges_key=args.charges_key,
keep_isolated_atoms=args.keep_isolated_atoms,
collections=collections,
avg_num_neighbors=model_foundation.interactions[0].avg_num_neighbors,
compute_avg_num_neighbors=False,
)
head_config_pt.collections = collections
head_configs.append(head_config_pt)
logging.info(
f"Total number of configurations: train={len(collections.train)}, valid={len(collections.valid)}"
)
# Atomic number table
# yapf: disable
for head_config in head_configs:
if head_config.atomic_numbers is None:
assert head_config.train_file.endswith(".xyz"), "Must specify atomic_numbers when using .h5 train_file input"
z_table_head = tools.get_atomic_number_table_from_zs(
z
for configs in (head_config.collections.train, head_config.collections.valid)
for config in configs
for z in config.atomic_numbers
)
head_config.atomic_numbers = z_table_head.zs
head_config.z_table = z_table_head
else:
if head_config.statistics_file is None:
logging.info("Using atomic numbers from command line argument")
else:
logging.info("Using atomic numbers from statistics file")
zs_list = ast.literal_eval(head_config.atomic_numbers)
assert isinstance(zs_list, list)
z_table_head = tools.AtomicNumberTable(zs_list)
head_config.atomic_numbers = zs_list
head_config.z_table = z_table_head
# yapf: enable
all_atomic_numbers = set()
for head_config in head_configs:
all_atomic_numbers.update(head_config.atomic_numbers)
z_table = AtomicNumberTable(sorted(list(all_atomic_numbers)))
logging.info(f"Atomic Numbers used: {z_table.zs}")
# Atomic energies
atomic_energies_dict = {}
for head_config in head_configs:
if head_config.atomic_energies_dict is None or len(head_config.atomic_energies_dict) == 0:
if head_config.train_file.endswith(".xyz") and head_config.E0s.lower() != "foundation":
atomic_energies_dict[head_config.head_name] = get_atomic_energies(
head_config.E0s, head_config.collections.train, head_config.z_table
)
elif head_config.E0s.lower() == "foundation":
assert args.foundation_model is not None
z_table_foundation = AtomicNumberTable(
[int(z) for z in model_foundation.atomic_numbers]
)
atomic_energies_dict[head_config.head_name] = {
z: model_foundation.atomic_energies_fn.atomic_energies[
z_table_foundation.z_to_index(z)
].item()
for z in z_table.zs
}
else:
atomic_energies_dict[head_config.head_name] = get_atomic_energies(head_config.E0s, None, head_config.z_table)
else:
atomic_energies_dict[head_config.head_name] = head_config.atomic_energies_dict
# Atomic energies for multiheads finetuning
if args.multiheads_finetuning:
assert (
model_foundation is not None
), "Model foundation must be provided for multiheads finetuning"
z_table_foundation = AtomicNumberTable(
[int(z) for z in model_foundation.atomic_numbers]
)
atomic_energies_dict["pt_head"] = {
z: model_foundation.atomic_energies_fn.atomic_energies[
z_table_foundation.z_to_index(z)
].item()
for z in z_table.zs
}
if args.model == "AtomicDipolesMACE":
atomic_energies = None
dipole_only = True
args.compute_dipole = True
args.compute_energy = False
args.compute_forces = False
args.compute_virials = False
args.compute_stress = False
else:
dipole_only = False
if args.model == "EnergyDipolesMACE":
args.compute_dipole = True
args.compute_energy = True
args.compute_forces = True
args.compute_virials = False
args.compute_stress = False
else:
args.compute_energy = True
args.compute_dipole = False
# atomic_energies: np.ndarray = np.array(
# [atomic_energies_dict[z] for z in z_table.zs]
# )
atomic_energies = dict_to_array(atomic_energies_dict, heads)
for head_config in head_configs:
logging.info(f"Atomic Energies used (z: eV) for head {head_config.head_name}: " + "{" + ", ".join([f"{z}: {atomic_energies_dict[head_config.head_name][z]}" for z in head_config.z_table.zs]) + "}")
valid_sets = {head: [] for head in heads}
train_sets = {head: [] for head in heads}
for head_config in head_configs:
if head_config.train_file.endswith(".xyz"):
train_sets[head_config.head_name] = [
data.AtomicData.from_config(
config, z_table=z_table, cutoff=args.r_max, heads=heads
)
for config in head_config.collections.train
]
valid_sets[head_config.head_name] = [
data.AtomicData.from_config(
config, z_table=z_table, cutoff=args.r_max, heads=heads
)
for config in head_config.collections.valid
]
elif head_config.train_file.endswith(".h5"):
train_sets[head_config.head_name] = data.HDF5Dataset(
head_config.train_file, r_max=args.r_max, z_table=z_table, heads=heads, head=head_config.head_name
)
valid_sets[head_config.head_name] = data.HDF5Dataset(
head_config.valid_file, r_max=args.r_max, z_table=z_table, heads=heads, head=head_config.head_name
)
else: # This case would be for when the file path is to a directory of multiple .h5 files
train_sets[head_config.head_name] = data.dataset_from_sharded_hdf5(
head_config.train_file, r_max=args.r_max, z_table=z_table, heads=heads, head=head_config.head_name
)
valid_sets[head_config.head_name] = data.dataset_from_sharded_hdf5(
head_config.valid_file, r_max=args.r_max, z_table=z_table, heads=heads, head=head_config.head_name
)
train_loader_head = torch_geometric.dataloader.DataLoader(
dataset=train_sets[head_config.head_name],
batch_size=args.batch_size,
shuffle=True,
drop_last=True,
pin_memory=args.pin_memory,
num_workers=args.num_workers,
generator=torch.Generator().manual_seed(args.seed),
)
head_config.train_loader = train_loader_head
# concatenate all the trainsets
train_set = ConcatDataset([train_sets[head] for head in heads])
train_sampler, valid_sampler = None, None
if args.distributed:
train_sampler = torch.utils.data.distributed.DistributedSampler(
train_set,
num_replicas=world_size,
rank=rank,
shuffle=True,
drop_last=True,
seed=args.seed,
)
valid_samplers = {}
for head, valid_set in valid_sets.items():
valid_sampler = torch.utils.data.distributed.DistributedSampler(
valid_set,
num_replicas=world_size,
rank=rank,
shuffle=True,
drop_last=True,
seed=args.seed,
)
valid_samplers[head] = valid_sampler
train_loader = torch_geometric.dataloader.DataLoader(
dataset=train_set,
batch_size=args.batch_size,
sampler=train_sampler,
shuffle=(train_sampler is None),
drop_last=(train_sampler is None),
pin_memory=args.pin_memory,
num_workers=args.num_workers,
generator=torch.Generator().manual_seed(args.seed),
)
valid_loaders = {heads[i]: None for i in range(len(heads))}
if not isinstance(valid_sets, dict):
valid_sets = {"Default": valid_sets}
for head, valid_set in valid_sets.items():
valid_loaders[head] = torch_geometric.dataloader.DataLoader(
dataset=valid_set,
batch_size=args.valid_batch_size,
sampler=valid_samplers[head] if args.distributed else None,
shuffle=False,
drop_last=False,
pin_memory=args.pin_memory,
num_workers=args.num_workers,
generator=torch.Generator().manual_seed(args.seed),
)
loss_fn = get_loss_fn(args, dipole_only, args.compute_dipole)
args.avg_num_neighbors = get_avg_num_neighbors(head_configs, args, train_loader, device)
# Model
model, output_args = configure_model(args, train_loader, atomic_energies, model_foundation, heads, z_table)
model.to(device)
logging.debug(model)
logging.info(f"Total number of parameters: {tools.count_parameters(model)}")
logging.info("")
logging.info("===========OPTIMIZER INFORMATION===========")
logging.info(f"Using {args.optimizer.upper()} as parameter optimizer")
logging.info(f"Batch size: {args.batch_size}")
if args.ema:
logging.info(f"Using Exponential Moving Average with decay: {args.ema_decay}")
logging.info(
f"Number of gradient updates: {int(args.max_num_epochs*len(train_set)/args.batch_size)}"
)
logging.info(f"Learning rate: {args.lr}, weight decay: {args.weight_decay}")
logging.info(loss_fn)
# Optimizer
param_options = get_params_options(args, model)
optimizer: torch.optim.Optimizer
optimizer = get_optimizer(args, param_options)
if args.device == "xpu":
logging.info("Optimzing model and optimzier for XPU")
model, optimizer = ipex.optimize(model, optimizer=optimizer)
logger = tools.MetricsLogger(
directory=args.results_dir, tag=tag + "_train"
) # pylint: disable=E1123
lr_scheduler = LRScheduler(optimizer, args)
swa: Optional[tools.SWAContainer] = None
swas = [False]
if args.swa:
swa, swas = get_swa(args, model, optimizer, swas, dipole_only)
checkpoint_handler = tools.CheckpointHandler(
directory=args.checkpoints_dir,
tag=tag,
keep=args.keep_checkpoints,
swa_start=args.start_swa,
)
start_epoch = 0
if args.restart_latest:
try:
opt_start_epoch = checkpoint_handler.load_latest(
state=tools.CheckpointState(model, optimizer, lr_scheduler),
swa=True,
device=device,
)
except Exception: # pylint: disable=W0703
opt_start_epoch = checkpoint_handler.load_latest(
state=tools.CheckpointState(model, optimizer, lr_scheduler),
swa=False,
device=device,
)
if opt_start_epoch is not None:
start_epoch = opt_start_epoch
ema: Optional[ExponentialMovingAverage] = None
if args.ema:
ema = ExponentialMovingAverage(model.parameters(), decay=args.ema_decay)
else:
for group in optimizer.param_groups:
group["lr"] = args.lr
if args.wandb:
setup_wandb(args)
if args.distributed:
distributed_model = DDP(model, device_ids=[local_rank])
else:
distributed_model = None
tools.train(
model=model,
loss_fn=loss_fn,
train_loader=train_loader,
valid_loaders=valid_loaders,
optimizer=optimizer,
lr_scheduler=lr_scheduler,
checkpoint_handler=checkpoint_handler,
eval_interval=args.eval_interval,
start_epoch=start_epoch,
max_num_epochs=args.max_num_epochs,
logger=logger,
patience=args.patience,
save_all_checkpoints=args.save_all_checkpoints,
output_args=output_args,
device=device,
swa=swa,
ema=ema,
max_grad_norm=args.clip_grad,
log_errors=args.error_table,
log_wandb=args.wandb,
distributed=args.distributed,
distributed_model=distributed_model,
train_sampler=train_sampler,
rank=rank,
)
logging.info("")
logging.info("===========RESULTS===========")
logging.info("Computing metrics for training, validation, and test sets")
train_valid_data_loader = {}
for head_config in head_configs:
data_loader_name = "train_" + head_config.head_name
train_valid_data_loader[data_loader_name] = head_config.train_loader
for head, valid_loader in valid_loaders.items():
data_load_name = "valid_" + head
train_valid_data_loader[data_load_name] = valid_loader
test_sets = {}
stop_first_test = False
test_data_loader = {}
if all(
head_config.test_file == head_configs[0].test_file
for head_config in head_configs
) and head_configs[0].test_file is not None:
stop_first_test = True
if all(
head_config.test_dir == head_configs[0].test_dir
for head_config in head_configs
) and head_configs[0].test_dir is not None:
stop_first_test = True
for head_config in head_configs:
if head_config.train_file.endswith(".xyz"):
print(head_config.test_file)
for name, subset in head_config.collections.tests:
print(name)
test_sets[name] = [
data.AtomicData.from_config(
config, z_table=z_table, cutoff=args.r_max, heads=heads
)
for config in subset
]
if head_config.test_dir is not None:
if not args.multi_processed_test:
test_files = get_files_with_suffix(head_config.test_dir, "_test.h5")
for test_file in test_files:
name = os.path.splitext(os.path.basename(test_file))[0]
test_sets[name] = data.HDF5Dataset(
test_file, r_max=args.r_max, z_table=z_table, heads=heads, head=head_config.head_name
)
else:
test_folders = glob(head_config.test_dir + "/*")
for folder in test_folders:
name = os.path.splitext(os.path.basename(test_file))[0]
test_sets[name] = data.dataset_from_sharded_hdf5(
folder, r_max=args.r_max, z_table=z_table, heads=heads, head=head_config.head_name
)
for test_name, test_set in test_sets.items():
print(test_name)
test_sampler = None
if args.distributed:
test_sampler = torch.utils.data.distributed.DistributedSampler(
test_set,
num_replicas=world_size,
rank=rank,
shuffle=True,
drop_last=True,
seed=args.seed,
)
try:
drop_last = test_set.drop_last
except AttributeError as e: # pylint: disable=W0612
drop_last = False
test_loader = torch_geometric.dataloader.DataLoader(
test_set,
batch_size=args.valid_batch_size,
shuffle=(test_sampler is None),
drop_last=drop_last,
num_workers=args.num_workers,
pin_memory=args.pin_memory,
)
test_data_loader[test_name] = test_loader
if stop_first_test:
break
for swa_eval in swas:
epoch = checkpoint_handler.load_latest(
state=tools.CheckpointState(model, optimizer, lr_scheduler),
swa=swa_eval,
device=device,
)
model.to(device)
if args.distributed:
distributed_model = DDP(model, device_ids=[local_rank])
model_to_evaluate = model if not args.distributed else distributed_model
if swa_eval:
logging.info(f"Loaded Stage two model from epoch {epoch} for evaluation")
else:
logging.info(f"Loaded Stage one model from epoch {epoch} for evaluation")
for param in model.parameters():
param.requires_grad = False
table_train_valid = create_error_table(
table_type=args.error_table,
all_data_loaders=train_valid_data_loader,
model=model_to_evaluate,
loss_fn=loss_fn,
output_args=output_args,
log_wandb=args.wandb,
device=device,
distributed=args.distributed,
)
logging.info("Error-table on TRAIN and VALID:\n" + str(table_train_valid))
if test_data_loader:
table_test = create_error_table(
table_type=args.error_table,
all_data_loaders=test_data_loader,
model=model_to_evaluate,
loss_fn=loss_fn,
output_args=output_args,
log_wandb=args.wandb,
device=device,
distributed=args.distributed,
)
logging.info("Error-table on TEST:\n" + str(table_test))
if rank == 0:
# Save entire model
if swa_eval:
model_path = Path(args.checkpoints_dir) / (tag + "_stagetwo.model")
else:
model_path = Path(args.checkpoints_dir) / (tag + ".model")
logging.info(f"Saving model to {model_path}")
if args.save_cpu:
model = model.to("cpu")
torch.save(model, model_path)
extra_files = {
"commit.txt": commit.encode("utf-8") if commit is not None else b"",
"config.yaml": json.dumps(
convert_to_json_format(extract_config_mace_model(model))
),
}
if swa_eval:
torch.save(
model, Path(args.model_dir) / (args.name + "_stagetwo.model")
)
try:
path_complied = Path(args.model_dir) / (
args.name + "_stagetwo_compiled.model"
)
logging.info(f"Compiling model, saving metadata {path_complied}")
model_compiled = jit.compile(deepcopy(model))
torch.jit.save(
model_compiled,
path_complied,
_extra_files=extra_files,
)
except Exception as e: # pylint: disable=W0703
pass
else:
torch.save(model, Path(args.model_dir) / (args.name + ".model"))
try:
path_complied = Path(args.model_dir) / (
args.name + "_compiled.model"
)
logging.info(f"Compiling model, saving metadata to {path_complied}")
model_compiled = jit.compile(deepcopy(model))
torch.jit.save(
model_compiled,
path_complied,
_extra_files=extra_files,
)
except Exception as e: # pylint: disable=W0703
pass
if args.distributed:
torch.distributed.barrier()
logging.info("Done")
if args.distributed:
torch.distributed.destroy_process_group()
if __name__ == "__main__":
main()