-
Notifications
You must be signed in to change notification settings - Fork 4
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
bug fixes and customization for analytical gradients
- Loading branch information
1 parent
8b29d02
commit 0747817
Showing
13 changed files
with
1,671 additions
and
439 deletions.
There are no files selected for viewing
246 changes: 246 additions & 0 deletions
246
docs/source/examples/.ipynb_checkpoints/multi_task_test_notebook-checkpoint.ipynb
Large diffs are not rendered by default.
Oops, something went wrong.
548 changes: 548 additions & 0 deletions
548
docs/source/examples/.ipynb_checkpoints/single_task_test_notebook-checkpoint.ipynb
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
322 changes: 322 additions & 0 deletions
322
examples/.ipynb_checkpoints/single_task_test_notebook_2d-checkpoint.ipynb
Large diffs are not rendered by default.
Oops, something went wrong.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,31 @@ | ||
#!/bin/bash | ||
#SBATCH -A m4055_g | ||
#SBATCH -C gpu | ||
#SBATCH -q regular | ||
#SBATCH -t 6:00:00 | ||
#SBATCH -n 256 | ||
#SBATCH --ntasks-per-node=4 | ||
#SBATCH -c 32 ##### 2 * [64/ntasks-per-node] | ||
#SBATCH --gpus-per-task=1 | ||
#SBATCH --gpu-bind=map_gpu:0,1,2,3 | ||
|
||
|
||
export SLURM_CPU_BIND="cores" | ||
number_of_workers=256 | ||
|
||
source /global/homes/m/mcn/gp2Scale/gp2Scale_env/bin/activate | ||
|
||
export OMP_NUM_THREADS=8 | ||
echo We have nodes: ${SLURM_JOB_NODELIST} | ||
|
||
echo "$SDN_IP_ADDR" | ||
|
||
hn=$(hostname -s) | ||
port="8786" | ||
echo ${port} | ||
echo "starting scheduler" | ||
dask-scheduler --no-dashboard --no-bokeh --no-show --host ${hn} --port ${port} & | ||
echo "starting workers" | ||
srun -o dask_worker_info.txt dask-worker ${hn}:${port} & | ||
echo "starting gp2Scale" | ||
python -u run_GPU.py ${hn}:${port} ${number_of_workers} |
Large diffs are not rendered by default.
Oops, something went wrong.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,234 @@ | ||
from dask.distributed import Client | ||
import socket | ||
import time | ||
import numpy | ||
from fvgp import gp | ||
import numpy as np | ||
import time | ||
from fvgp.gp2Scale import gp2Scale | ||
import argparse | ||
import datetime | ||
import time | ||
import sys | ||
import torch | ||
from dask.distributed import performance_report | ||
import dask.distributed as distributed | ||
|
||
|
||
def client_is_ready(ip, n_workers): | ||
ready = False | ||
print("checking the client", flush = True) | ||
counter = 0 | ||
while ready is False: | ||
try: | ||
c = Client(ip) | ||
n = len(list(c.scheduler_info()["workers"])) | ||
if n >= n_workers: | ||
ready=True | ||
return ready | ||
else: | ||
print("only ",n," of desired",n_workers," workers available", flush = True) | ||
except: pass | ||
time.sleep(5) | ||
counter += 1 | ||
if counter > 20: print("getting the client is taking a long time: ", counter * 5, "seconds", flush = True) | ||
|
||
|
||
def kernel_gpu(x1,x2, hps,obj): | ||
|
||
def b(x,x0, r, ampl, device): | ||
""" | ||
evaluates the bump function | ||
x ... a point (1d numpy array) | ||
x0 ... 1d numpy array of location of bump function | ||
returns the bump function b(x,x0) with radius r | ||
""" | ||
x_new = x - x0 | ||
d = torch.linalg.norm(x_new, axis = 1) | ||
a = torch.zeros(d.shape).to(device, dtype = torch.float32) | ||
a = 1.0 - (d**2/r**2) | ||
i = torch.where(a > 0.0) | ||
bump = torch.zeros(a.shape).to(device, dtype = torch.float32) | ||
e = torch.exp((-1.0/a[i])+1).to(device, dtype = torch.float32) | ||
bump[i] = ampl * e | ||
return bump | ||
|
||
|
||
def f(x,x0, radii, amplts, device): | ||
b1 = b(x, x0[0:3], radii[0], amplts[0], device) ###x0[0] ... D-dim location of bump func 1 | ||
b2 = b(x, x0[3:6], radii[1], amplts[1], device) ###x0[1] ... D-dim location of bump func 2 | ||
b3 = b(x, x0[6:9], radii[2], amplts[2], device) ###x0[1] ... D-dim location of bump func 2 | ||
b4 = b(x, x0[9:12],radii[3], amplts[3], device) ###x0[1] ... D-dim location of bump func 2 | ||
return b1 + b2 + b3 + b4 | ||
|
||
def g(x,x0, radii, amplts,device): | ||
b1 = b(x, x0[0:3], radii[0], amplts[0], device) ###x0[0] ... D-dim location of bump func 1 | ||
b2 = b(x, x0[3:6], radii[1], amplts[1], device) ###x0[1] ... D-dim location of bump func 2 | ||
b3 = b(x, x0[6:9], radii[2], amplts[2], device) ###x0[1] ... D-dim location of bump func 2 | ||
b4 = b(x, x0[9:12],radii[3], amplts[3], device) ###x0[1] ... D-dim location of bump func 2 | ||
return b1 + b2 + b3 + b4 | ||
|
||
def get_distance_matrix(x1,x2,device): | ||
d = torch.zeros((len(x1),len(x2))).to(device, dtype = torch.float32) | ||
for i in range(x1.shape[1]): | ||
d += ((x1[:,i].reshape(-1, 1) - x2[:,i]))**2 | ||
return torch.sqrt(d) | ||
|
||
|
||
def sparse_stat_kernel(x1,x2, hps,device): | ||
d = get_distance_matrix(x1,x2,device) | ||
d[d == 0.0] = 1e-6 | ||
d[d > hps] = hps | ||
|
||
d_hps = d/hps | ||
d_hpss= d_hps**2 | ||
sq = torch.sqrt(1.0 - d_hpss) | ||
|
||
|
||
kernel = (torch.sqrt(torch.tensor(2.0))/(3.0*torch.sqrt(torch.tensor(3.141592653))))*\ | ||
((3.0*d_hpss * torch.log((d_hps)/(1+sq)))+\ | ||
((2.0*d_hpss + 1.0)*sq)) | ||
return kernel | ||
|
||
|
||
def ks_gpu(x1,x2,hps,cuda_device): | ||
k1 = torch.outer(f(x1,hps[0:12],hps[12:16],hps[16:20],cuda_device), | ||
f(x2,hps[0:12],hps[12:16],hps[16:20],cuda_device)) + \ | ||
torch.outer(g(x1,hps[20:32],hps[32:36],hps[36:40],cuda_device), | ||
g(x2,hps[20:32],hps[32:36],hps[36:40],cuda_device)) | ||
k2 = sparse_stat_kernel(x1,x2, hps[41],cuda_device) | ||
return k1 + hps[40]*k2 | ||
|
||
cuda_device = torch.device("cuda:0") | ||
x1_dev = torch.from_numpy(x1).to(cuda_device, dtype = torch.float32) | ||
x2_dev = torch.from_numpy(x2).to(cuda_device, dtype = torch.float32) | ||
hps_dev = torch.from_numpy(hps).to(cuda_device, dtype = torch.float32) | ||
ksparse = ks_gpu(x1_dev,x2_dev,hps_dev,cuda_device).cpu().numpy() | ||
return ksparse | ||
|
||
|
||
def normalize(v): | ||
v = v - np.min(v) | ||
v = v/np.max(v) | ||
return v | ||
|
||
|
||
def main(): | ||
start_time = time.time() | ||
print("inputs to the run script: ",sys.argv, flush = True) | ||
print("port: ", str(sys.argv[1]), flush = True) | ||
if client_is_ready(str(sys.argv[1]),int(sys.argv[2])): | ||
client = Client(str(sys.argv[1])) | ||
print("Client is ready", flush = True) | ||
print(datetime.datetime.now().isoformat()) | ||
print("client received: ", client, flush = True) | ||
|
||
print("Everything is ready to call gp2Scale after ", time.time() - start_time, flush = True) | ||
print("Number of GPUs: ", torch.cuda.device_count()) | ||
target_worker_count = int(sys.argv[2]) | ||
with performance_report(filename="dask-report-gpu.html"): | ||
input_dim = 3 | ||
target_worker_count = int(sys.argv[2]) | ||
|
||
station_locations = np.load("station_coord.npy") | ||
temperatures = np.load("data.npy") | ||
N = len(station_locations) * len(temperatures) | ||
x_data = np.zeros((N,3)) | ||
y_data = np.zeros((N)) | ||
count = 0 | ||
for i in range(len(temperatures)): | ||
for j in range(len(temperatures[0])): | ||
x_data[count] = np.array([station_locations[j,0],station_locations[j,1],float(i)]) | ||
y_data[count] = temperatures[i,j] | ||
count += 1 | ||
|
||
non_nan_indices = np.where(y_data == y_data) ###nans in data | ||
x_data = x_data[non_nan_indices] | ||
y_data = y_data[non_nan_indices] | ||
x_data = x_data[::10] ##1000: about 50 000 points; 100: 500 000; 10: 5 million | ||
y_data = y_data[::10] | ||
x_data[:,0] = normalize(x_data[:,0]) | ||
x_data[:,1] = normalize(x_data[:,1]) | ||
x_data[:,2] = normalize(x_data[:,2]) | ||
print(np.min(x_data[:,0]),np.max(x_data[:,0])) | ||
print(np.min(x_data[:,1]),np.max(x_data[:,1])) | ||
print(np.min(x_data[:,2]),np.max(x_data[:,2])) | ||
|
||
N = len(x_data) | ||
|
||
hps_n = 42 | ||
|
||
hps_bounds = np.array([ | ||
[0.,1.], ##pos bump 1 f comp 1 | ||
[0.,1.], ##pos bump 1 f comp 2 | ||
[0.,1.], ##pos bump 1 f comp 3 | ||
# | ||
[0.,1.], ##pos bump 2 f | ||
[0.,1.], ##pos bump 2 f | ||
[0.,1.], ##pos bump 2 f | ||
# | ||
[0.,1.], ##pos bump 3 f | ||
[0.,1.], ##pos bump 3 f | ||
[0.,1.], ##pos bump 3 f | ||
# | ||
[0.,1.], ##pos bump 4 f | ||
[0.,1.], ##pos bump 4 f | ||
[0.,1.], ##pos bump 4 f | ||
# | ||
[0.01,0.05], ##radius bump 1 f | ||
[0.01,0.05], ##...2 | ||
[0.01,0.05], ##...3 | ||
[0.01,0.05], ##...4 | ||
[0.1,1.], ##ampl bump 1 f | ||
[0.1,1.], ##...2 | ||
[0.1,1.], ##...3 | ||
[0.1,1.], ##...4 | ||
# | ||
[0.,1.], ##pos bump 1 g comp 1 | ||
[0.,1.], ##pos bump 1 g comp 2 | ||
[0.,1.], ##pos bump 1 g comp 3 | ||
# | ||
[0.,1.], ##pos bump 2 g comp 1 | ||
[0.,1.], ##pos bump 2 g comp 2 | ||
[0.,1.], ##pos bump 2 g comp 3 | ||
# | ||
[0.,1.], ##pos bump 3 g comp 1 | ||
[0.,1.], ##pos bump 3 g comp 2 | ||
[0.,1.], ##pos bump 3 g comp 3 | ||
# | ||
[0.,1.], ##pos bump 4 g comp 1 | ||
[0.,1.], ##pos bump 4 g comp 2 | ||
[0.,1.], ##pos bump 4 g comp 3 | ||
# | ||
[0.01,0.05], ##radius bump 1 g | ||
[0.01,0.05], ##...2 | ||
[0.01,0.05], ##...3 | ||
[0.01,0.05], ##...4 | ||
[0.1,1.], ##ampl bump 1 g | ||
[0.1,1.], ##...2 | ||
[0.1,1.], ##...3 | ||
[0.1,1.], ##...4 | ||
[0.1,10.], ##signal var of stat kernel | ||
[0.01,0.02] ##length scale for stat kernel | ||
]) | ||
#init_hps = np.random.uniform(size = len(hps_bounds), low = hps_bounds[:,0], high = hps_bounds[:,1]) | ||
init_hps = np.load("latest_hps.npy", allow_pickle = True) | ||
print(init_hps) | ||
print(hps_bounds) | ||
print("INITIALIZED") | ||
st = time.time() | ||
my_gp = gp2Scale(input_dim, x_data, y_data, init_hps, 10000, target_worker_count, | ||
gp_kernel_function = kernel_gpu, info = False, | ||
covariance_dask_client = client) | ||
|
||
print("init done after: ",time.time() - st," seconds") | ||
print("===============") | ||
#print("Log Likelihood: ", my_gp.log_likelihood(my_gp.hyperparameters)) | ||
#print("done after: ",time.time() - st," seconds") | ||
|
||
my_gp.train(hps_bounds, max_iter = 100, init_hyperparameters = init_hps) | ||
|
||
|
||
if __name__ == '__main__': | ||
main() | ||
|
Large diffs are not rendered by default.
Oops, something went wrong.
Large diffs are not rendered by default.
Oops, something went wrong.
Oops, something went wrong.