-
Notifications
You must be signed in to change notification settings - Fork 5
/
dataset_train_graph.py
42 lines (32 loc) · 1.58 KB
/
dataset_train_graph.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
from sklearn.model_selection import KFold
from torch.utils.data import Dataset
import numpy as np
import glob
def default_transform(x):
return x
class COPD_dataset(Dataset):
def __init__(self, stage, cfg, transforms=default_transform, graph_cutoff=0.13):
self.cfg = cfg
self.root_dir = cfg.root_dir
self.transforms = transforms
self.graph_cutoff = graph_cutoff # hyper-parameter used to control graph sparsity
self.sid_list = []
for item in glob.glob(self.cfg.root_dir+"patch/"+"*_patch.npy"):
self.sid_list.append(item.split('/')[-1][:-10])
self.sid_list.sort()
self.patch_loc = np.load("/pghbio/dbmi/batmanlab/lisun/copd/gnn_shared/data/patch_data_32_6_reg/19676E_INSP_STD_JHU_COPD_BSpline_Iso1_patch_loc.npy")
self.patch_loc = self.patch_loc / self.patch_loc.max(0) # column-wise norm
self.sid_list = np.asarray(self.sid_list)
print(stage+" dataset size:", len(self))
def __len__(self):
return len(self.sid_list)
def __getitem__(self, idx):
img = np.load(self.root_dir+"patch/"+self.sid_list[idx]+"_patch.npy")
img = img + 1024.
# Get 2 augmented images (positive pair)
img = self.transforms(img)
img[0] = img[0][:,None,:,:,:]/632.-1 # Normalize to [-1,1], 632=(1024+240)/2
img[1] = img[1][:,None,:,:,:]/632.-1 # Normalize to [-1,1], 632=(1024+240)/2
adj = np.load(self.root_dir+"adj/"+self.sid_list[idx]+"_adj.npy")
adj=(adj>self.graph_cutoff).astype(np.int)
return img, self.patch_loc.copy(), adj