Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update gromacs model devi engine #493

Merged
merged 15 commits into from
Aug 14, 2021
Merged
Show file tree
Hide file tree
Changes from 11 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
7 changes: 6 additions & 1 deletion dpgen/generator/lib/gaussian.py
Original file line number Diff line number Diff line change
Expand Up @@ -114,8 +114,13 @@ def make_gaussian_input(sys_data, fp_params):
keywords = [keywords]
else:
keywords = keywords.copy()

# assume default charge is zero and default spin multiplicity is 1
charge = fp_params.get('charge', 0)
if 'charge' in sys_data.keys():
charge = sys_data['charge']
else:
charge = fp_params.get('charge', 0)

use_fragment_guesses = False
multiplicity = fp_params.get('multiplicity', 'auto')
if type(multiplicity) == int:
Expand Down
163 changes: 106 additions & 57 deletions dpgen/generator/run.py
Original file line number Diff line number Diff line change
Expand Up @@ -183,9 +183,12 @@ def dump_to_poscar(dump, poscar, type_map, fmt = "lammps/dump") :
sys = dpdata.System(dump, fmt = fmt, type_map = type_map)
sys.to_vasp_poscar(poscar)

def dump_to_deepmd_raw(dump, deepmd_raw, type_map, fmt='gromacs/gro'):
def dump_to_deepmd_raw(dump, deepmd_raw, type_map, fmt='gromacs/gro', charge=None):
system = dpdata.System(dump, fmt = fmt, type_map = type_map)
system.to_deepmd_raw(deepmd_raw)
if charge is not None:
with open(os.path.join(deepmd_raw, "charge"), 'w') as f:
f.write(str(charge))


def make_train (iter_index,
Expand All @@ -205,7 +208,14 @@ def make_train (iter_index,
training_init_model = jdata.get('training_init_model', False)
training_reuse_iter = jdata.get('training_reuse_iter')
training_reuse_old_ratio = jdata.get('training_reuse_old_ratio', None)
training_reuse_stop_batch = jdata.get('training_reuse_stop_batch', 400000)

if 'training_reuse_stop_batch' in jdata.keys():
training_reuse_stop_batch = jdata['training_reuse_stop_batch']
elif 'training_reuse_numb_steps' in jdata.keys():
training_reuse_stop_batch = jdata['training_reuse_numb_steps']
AnguseZhang marked this conversation as resolved.
Show resolved Hide resolved
else:
training_reuse_stop_batch = 40000
AnguseZhang marked this conversation as resolved.
Show resolved Hide resolved

training_reuse_start_lr = jdata.get('training_reuse_start_lr', 1e-4)
training_reuse_start_pref_e = jdata.get('training_reuse_start_pref_e', 0.1)
training_reuse_start_pref_f = jdata.get('training_reuse_start_pref_f', 100)
Expand Down Expand Up @@ -346,10 +356,12 @@ def make_train (iter_index,
# set training reuse model
if training_reuse_iter is not None and iter_index >= training_reuse_iter:
if LooseVersion('1') <= LooseVersion(mdata["deepmd_version"]) < LooseVersion('2'):
jinput['training']['stop_batch'] = training_reuse_stop_batch
jinput['training']['auto_prob_style'] \
="prob_sys_size; 0:%d:%f; %d:%d:%f" \
%(old_range, training_reuse_old_ratio, old_range, len(init_data_sys), 1.-training_reuse_old_ratio)
elif LooseVersion('2') <= LooseVersion(mdata["deepmd_version"]) < LooseVersion('3'):
jinput['training']['numb_steps'] = training_reuse_stop_batch
jinput['training']['training_data']['auto_prob'] \
="prob_sys_size; 0:%d:%f; %d:%d:%f" \
%(old_range, training_reuse_old_ratio, old_range, len(init_data_sys), 1.-training_reuse_old_ratio)
Expand All @@ -360,7 +372,7 @@ def make_train (iter_index,
if jinput['loss'].get('start_pref_f') is not None:
jinput['loss']['start_pref_f'] = training_reuse_start_pref_f
jinput['learning_rate']['start_lr'] = training_reuse_start_lr
jinput['training']['stop_batch'] = training_reuse_stop_batch


for ii in range(numb_models) :
task_path = os.path.join(work_path, train_task_fmt % ii)
Expand Down Expand Up @@ -1076,6 +1088,9 @@ def _make_model_devi_native(iter_index, jdata, mdata, conf_systems):
sys_counter += 1

def _make_model_devi_native_gromacs(iter_index, jdata, mdata, conf_systems):
# only support for deepmd v2.0
if LooseVersion(mdata['deepmd_version']) < LooseVersion('2.0'):
raise RuntimeError("Only support deepmd-kit 2.x for model_devi_engine='gromacs'")
model_devi_jobs = jdata['model_devi_jobs']
if (iter_index >= len(model_devi_jobs)) :
return False
Expand All @@ -1086,6 +1101,13 @@ def _make_model_devi_native_gromacs(iter_index, jdata, mdata, conf_systems):
else:
model_devi_dt = jdata['model_devi_dt']
nsteps = cur_job.get("nsteps", None)
lambdas = cur_job.get("lambdas", [1.0])
temps = cur_job.get("temps", [298.0])

for ll in lambdas:
if ll > 1:
raise RuntimeError("lambda is larger than 1.0")
AnguseZhang marked this conversation as resolved.
Show resolved Hide resolved

if nsteps is None:
raise RuntimeError("nsteps is None, you should set nsteps in model_devi_jobs!")
# Currently Gromacs engine is not supported for different temperatures!
Expand All @@ -1111,48 +1133,50 @@ def _make_model_devi_native_gromacs(iter_index, jdata, mdata, conf_systems):
conf_counter = 0
task_counter = 0
for cc in ss :
task_name = make_model_devi_task_name(sys_idx[sys_counter], task_counter)
#conf_name = make_model_devi_conf_name(sys_idx[sys_counter], conf_counter) + '.lmp'
task_path = os.path.join(work_path, task_name)
# dlog.info(task_path)
create_path(task_path)
#create_path(os.path.join(task_path, 'traj'))
#loc_conf_name = 'conf.lmp'
gromacs_settings = jdata.get("gromacs_settings" , "")
for key,file in gromacs_settings.items():
if key != "traj_filename" and key != "mdp_filename":
os.symlink(os.path.join(cc,file), os.path.join(task_path, file))

# input.json for DP-Gromacs
with open(os.path.join(cc, "input.json")) as f:
input_json = json.load(f)
input_json["graph_file"] = models[0]
with open(os.path.join(task_path,'input.json'), 'w') as _outfile:
json.dump(input_json, _outfile, indent = 4)

# trj_freq
trj_freq = cur_job.get("trj_freq", 10)
mdp = MDP()
mdp.read(os.path.join(cc, gromacs_settings['mdp_filename']))
mdp['nstcomm'] = trj_freq
mdp['nstxout'] = trj_freq
mdp['nstlog'] = trj_freq
mdp['nstenergy'] = trj_freq
# dt
mdp['dt'] = dt
mdp.write(os.path.join(task_path, gromacs_settings['mdp_filename']))

cwd_ = os.getcwd()
os.chdir(task_path)
job = {}

job["model_devi_dt"] = model_devi_dt
job["nsteps"] = nsteps
with open('job.json', 'w') as _outfile:
json.dump(job, _outfile, indent = 4)
os.chdir(cwd_)

task_counter += 1
for ll in lambdas:
for tt in temps:
task_name = make_model_devi_task_name(sys_idx[sys_counter], task_counter)
task_path = os.path.join(work_path, task_name)
create_path(task_path)
gromacs_settings = jdata.get("gromacs_settings" , "")
for key,file in gromacs_settings.items():
if key != "traj_filename" and key != "mdp_filename" and key != "group_name":
os.symlink(os.path.join(cc,file), os.path.join(task_path, file))
# input.json for DP-Gromacs
with open(os.path.join(cc, "input.json")) as f:
input_json = json.load(f)
input_json["graph_file"] = models[0]
input_json["lambda"] = ll
with open(os.path.join(task_path,'input.json'), 'w') as _outfile:
json.dump(input_json, _outfile, indent = 4)

# trj_freq
trj_freq = cur_job.get("trj_freq", 10)
mdp = MDP()
mdp.read(os.path.join(cc, gromacs_settings['mdp_filename']))
mdp['nstcomm'] = trj_freq
mdp['nstxout'] = trj_freq
mdp['nstlog'] = trj_freq
mdp['nstenergy'] = trj_freq
# dt
mdp['dt'] = model_devi_dt
# temps
if "ref_t" in list(mdp.keys()):
mdp["ref_t"] = tt
else:
mdp["ref-t"] = tt
mdp.write(os.path.join(task_path, gromacs_settings['mdp_filename']))

cwd_ = os.getcwd()
os.chdir(task_path)
job = {}
job["trj_freq"] = cur_job["trj_freq"]
job["model_devi_dt"] = model_devi_dt
job["nsteps"] = nsteps
with open('job.json', 'w') as _outfile:
json.dump(job, _outfile, indent = 4)
os.chdir(cwd_)
task_counter += 1
conf_counter += 1
sys_counter += 1

Expand Down Expand Up @@ -1208,21 +1232,32 @@ def run_model_devi (iter_index,
if use_plm_path:
forward_files += ['plmpath.pdb']
elif model_devi_engine == "gromacs":

gromacs_settings = jdata.get("gromacs_settings", {})
mdp_filename = gromacs_settings.get("mdp_filename", "md.mdp")
topol_filename = gromacs_settings.get("topol_filename", "processed.top")
conf_filename = gromacs_settings.get("conf_filename", "conf.gro")
index_filename = gromacs_settings.get("index_filename", "index.raw")
# Initial reference to process pbc condition.
# Default is em.tpr
ref_filename = gromacs_settings.get("ref_filename", "em.tpr")
deffnm = gromacs_settings.get("deffnm", "deepmd")
maxwarn = gromacs_settings.get("maxwarn", 1)
traj_filename = gromacs_settings.get("traj_filename", "deepmd_traj.gro")
grp_name = gromacs_settings.get("group_name", "Other")
nsteps = cur_job["nsteps"]
trj_freq = cur_job.get("trj_freq", 10)

command = "%s grompp -f %s -p %s -c %s -o %s -maxwarn %d" % (model_devi_exec, mdp_filename, topol_filename, conf_filename, deffnm, maxwarn)
command += "&& %s mdrun -deffnm %s -nsteps %d" %(model_devi_exec, deffnm, nsteps)
command += "&& %s mdrun -deffnm %s -nsteps %d" %(model_devi_exec, deffnm, nsteps)
command += "&& echo -e \"%s\n%s\n\" | %s trjconv -s %s -f %s.trr -o %s -pbc mol -ur compact -center" % (grp_name, grp_name, model_devi_exec, ref_filename, deffnm, traj_filename)
command += "&& if [ ! -d traj ]; then \n mkdir traj; fi\n"
command += f"python -c \"import dpdata;system = dpdata.System('{traj_filename}', fmt='gromacs/gro'); [system.to_gromacs_gro('traj/%d.gromacstrj' % (i * {trj_freq}), frame_idx=i) for i in range(system.get_nframes())]; system.to_deepmd_npy('traj_deepmd')\""
command += "&& dp model-devi -m ../graph.000.pb ../graph.001.pb ../graph.002.pb ../graph.003.pb -s traj_deepmd -o model_devi.out"
commands = [command]
forward_files = [mdp_filename, topol_filename, conf_filename, index_filename, "input.json" ]
backward_files = ["%s.tpr" % deffnm, "%s.log" %deffnm , 'model_devi.out', 'model_devi.log']

forward_files = [mdp_filename, topol_filename, conf_filename, index_filename, ref_filename, "input.json", "job.json" ]
backward_files = ["%s.tpr" % deffnm, "%s.log" %deffnm , traj_filename, 'model_devi.out', "traj", "traj_deepmd" ]


cwd = os.getcwd()
Expand Down Expand Up @@ -1351,6 +1386,10 @@ def _make_fp_vasp_inner (modd_path,
system_index.sort()

fp_tasks = []

charges_recorder = [] # record charges for each fp_task
charges_map = jdata.get("sys_charges", [])

cluster_cutoff = jdata['cluster_cutoff'] if jdata.get('use_clusters', False) else None
# skip save *.out if detailed_report_make_fp is False, default is True
detailed_report_make_fp = jdata.get("detailed_report_make_fp", True)
Expand Down Expand Up @@ -1464,11 +1503,11 @@ def _make_fp_vasp_inner (modd_path,
continue

if fp_cluster_vacuum is not None:
assert fp_cluster_vacuum >0
skip_cluster = check_cluster(conf_name, fp_cluster_vacuum)
if skip_cluster:
count_bad_cluster +=1
continue
assert fp_cluster_vacuum >0
skip_cluster = check_cluster(conf_name, fp_cluster_vacuum)
if skip_cluster:
count_bad_cluster +=1
continue

# link job.json
job_name = os.path.join(tt, "job.json")
Expand All @@ -1484,6 +1523,8 @@ def _make_fp_vasp_inner (modd_path,
fp_task_path = os.path.join(work_path, fp_task_name)
create_path(fp_task_path)
fp_tasks.append(fp_task_path)
if charges_map:
charges_recorder.append(charges_map[int(ss)])
cwd = os.getcwd()
os.chdir(fp_task_path)
if cluster_cutoff is None:
Expand All @@ -1501,13 +1542,18 @@ def _make_fp_vasp_inner (modd_path,
dlog.info("system {0:s} skipped {1:6d} confs with bad cluster, {2:6d} remains".format(ss, count_bad_cluster, numb_task - count_bad_cluster))
if cluster_cutoff is None:
cwd = os.getcwd()
for ii in fp_tasks:
os.chdir(ii)
for idx, task in enumerate(fp_tasks):
os.chdir(task)
if model_devi_engine == "lammps":
dump_to_poscar('conf.dump', 'POSCAR', type_map, fmt = "lammps/dump")
if charges_map:
warnings.warn('"sys_charges" keyword only support for gromacs engine now.')
elif model_devi_engine == "gromacs":
# dump_to_poscar('conf.dump', 'POSCAR', type_map, fmt = "gromacs/gro")
dump_to_deepmd_raw('conf.dump', 'deepmd.raw', type_map, fmt = 'gromacs/gro')
if charges_map:
dump_to_deepmd_raw('conf.dump', 'deepmd.raw', type_map, fmt='gromacs/gro', charge=charges_recorder[idx])
else:
dump_to_deepmd_raw('conf.dump', 'deepmd.raw', type_map, fmt='gromacs/gro', charge=None)
else:
raise RuntimeError("unknown model_devi engine", model_devi_engine)
os.chdir(cwd)
Expand Down Expand Up @@ -1933,6 +1979,8 @@ def make_fp_gaussian(iter_index,
sys_data = dpdata.System('POSCAR').data
elif model_devi_engine == "gromacs":
sys_data = dpdata.System("deepmd.raw", fmt='deepmd/raw').data
if os.path.isfile('deepmd.raw/charge'):
sys_data['charge'] = int(np.loadtxt('deepmd.raw/charge', dtype=int))
ret = make_gaussian_input(sys_data, fp_params)
with open('input', 'w') as fp:
fp.write(ret)
Expand Down Expand Up @@ -2483,6 +2531,7 @@ def post_fp_gaussian (iter_index,
sys_output = glob.glob(os.path.join(work_path, "task.%s.*/output"%ss))
sys_output.sort()
for idx,oo in enumerate(sys_output) :
# TODO : UnboundLocalError sometimes occurs when parsing gaussian log
sys = dpdata.LabeledSystem(oo, fmt = 'gaussian/log')
if len(sys) > 0:
sys.check_type_map(type_map = jdata['type_map'])
Expand Down
4 changes: 2 additions & 2 deletions tests/generator/test_gromacs_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@ def _copy_outputs(self, path_1, path_2):
def test_make_model_devi_gromacs(self):
flag = make_model_devi(iter_index=0,
jdata=self.jdata,
mdata={})
mdata={"deepmd_version": "2.0"})
self.assertTrue(flag)
self.assertTrue(os.path.exists(self.model_devi_path))
self.assertTrue(os.path.exists(self.model_devi_task_path))
Expand All @@ -108,7 +108,7 @@ def test_make_fp_gaussian(self):


def tearDown(self):
#pass
# pass
shutil.rmtree(self.iter_path)
if __name__ == '__main__':
unittest.main()
Expand Down