hc99's picture
Add files using upload-large-folder tool
e4b9a7b verified
raw
history blame
11.1 kB
# Copyright 2020 MONAI Consortium
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
import time
from argparse import ArgumentParser
import numpy as np
import torch
from data_utils import get_filenames, load_data_and_mask
from torchgpipe import GPipe
from torchgpipe.balance import balance_by_size
from unet_pipe import UNetPipe, flatten_sequential
from monai.data import Dataset, list_data_collate
from monai.losses import DiceLoss, FocalLoss
from monai.metrics import compute_meandice
from monai.transforms import AddChannelDict, Compose, Rand3DElasticd, RandCropByPosNegLabeld, SpatialPadd
from monai.utils import first
N_CLASSES = 10
TRAIN_PATH = "./data/HaN/train/" # training data folder
VAL_PATH = "./data/HaN/test/" # validation data folder
torch.backends.cudnn.enabled = True
class ImageLabelDataset:
"""
Load image and multi-class labels based on the predefined folder structure.
"""
def __init__(self, path, n_class=10):
self.path = path
self.data = sorted(os.listdir(path))
self.n_class = n_class
def __getitem__(self, index):
data = os.path.join(self.path, self.data[index])
train_data, train_masks_data = get_filenames(data)
data = load_data_and_mask(train_data, train_masks_data) # read into a data dict
# loading image
data["image"] = data["image"].astype(np.float32) # shape (H W D)
# loading labels
class_shape = (1,) + data["image"].shape
mask0 = np.zeros(class_shape)
mask_list = []
flagvect = np.ones((self.n_class,), np.float32)
for i, mask in enumerate(data["label"]):
if mask is None:
mask = np.zeros(class_shape)
flagvect[0] = 0
flagvect[i + 1] = 0
mask0 = np.logical_or(mask0, mask)
mask_list.append(mask.reshape(class_shape))
mask0 = 1 - mask0
data["label"] = np.concatenate([mask0] + mask_list, axis=0).astype(np.uint8) # shape (C H W D)
# setting flags
data["with_complete_groundtruth"] = flagvect # flagvec is a boolean indicator for complete annotation
return data
def __len__(self):
return len(self.data)
def train(n_feat, crop_size, bs, ep, optimizer="rmsprop", lr=5e-4, pretrain=None):
model_name = f"./HaN_{n_feat}_{bs}_{ep}_{crop_size}_{lr}_"
print(f"save the best model as '{model_name}' during training.")
crop_size = [int(cz) for cz in crop_size.split(",")]
print(f"input image crop_size: {crop_size}")
# starting training set loader
train_images = ImageLabelDataset(path=TRAIN_PATH, n_class=N_CLASSES)
if np.any([cz == -1 for cz in crop_size]): # using full image
train_transform = Compose(
[
AddChannelDict(keys="image"),
Rand3DElasticd(
keys=("image", "label"),
spatial_size=crop_size,
sigma_range=(10, 50), # 30
magnitude_range=(600, 1200), # 1000
prob=0.8,
rotate_range=(np.pi / 12, np.pi / 12, np.pi / 12),
shear_range=(np.pi / 18, np.pi / 18, np.pi / 18),
translate_range=tuple(sz * 0.05 for sz in crop_size),
scale_range=(0.2, 0.2, 0.2),
mode=("bilinear", "nearest"),
padding_mode=("border", "zeros"),
),
]
)
train_dataset = Dataset(train_images, transform=train_transform)
# when bs > 1, the loader assumes that the full image sizes are the same across the dataset
train_dataloader = torch.utils.data.DataLoader(train_dataset, num_workers=4, batch_size=bs, shuffle=True)
else:
# draw balanced foreground/background window samples according to the ground truth label
train_transform = Compose(
[
AddChannelDict(keys="image"),
SpatialPadd(keys=("image", "label"), spatial_size=crop_size), # ensure image size >= crop_size
RandCropByPosNegLabeld(
keys=("image", "label"), label_key="label", spatial_size=crop_size, num_samples=bs
),
Rand3DElasticd(
keys=("image", "label"),
spatial_size=crop_size,
sigma_range=(10, 50), # 30
magnitude_range=(600, 1200), # 1000
prob=0.8,
rotate_range=(np.pi / 12, np.pi / 12, np.pi / 12),
shear_range=(np.pi / 18, np.pi / 18, np.pi / 18),
translate_range=tuple(sz * 0.05 for sz in crop_size),
scale_range=(0.2, 0.2, 0.2),
mode=("bilinear", "nearest"),
padding_mode=("border", "zeros"),
),
]
)
train_dataset = Dataset(train_images, transform=train_transform) # each dataset item is a list of windows
train_dataloader = torch.utils.data.DataLoader( # stack each dataset item into a single tensor
train_dataset, num_workers=4, batch_size=1, shuffle=True, collate_fn=list_data_collate
)
first_sample = first(train_dataloader)
print(first_sample["image"].shape)
# starting validation set loader
val_transform = Compose([AddChannelDict(keys="image")])
val_dataset = Dataset(ImageLabelDataset(VAL_PATH, n_class=N_CLASSES), transform=val_transform)
val_dataloader = torch.utils.data.DataLoader(val_dataset, num_workers=1, batch_size=1)
print(val_dataset[0]["image"].shape)
print(f"training images: {len(train_dataloader)}, validation images: {len(val_dataloader)}")
model = UNetPipe(spatial_dims=3, in_channels=1, out_channels=N_CLASSES, n_feat=n_feat)
model = flatten_sequential(model)
lossweight = torch.from_numpy(np.array([2.22, 1.31, 1.99, 1.13, 1.93, 1.93, 1.0, 1.0, 1.90, 1.98], np.float32))
if optimizer.lower() == "rmsprop":
optimizer = torch.optim.RMSprop(model.parameters(), lr=lr) # lr = 5e-4
elif optimizer.lower() == "momentum":
optimizer = torch.optim.SGD(model.parameters(), lr=lr, momentum=0.9) # lr = 1e-4 for finetuning
else:
raise ValueError(f"Unknown optimizer type {optimizer}. (options are 'rmsprop' and 'momentum').")
# config GPipe
x = first_sample["image"].float()
x = torch.autograd.Variable(x.cuda())
partitions = torch.cuda.device_count()
print(f"partition: {partitions}, input: {x.size()}")
balance = balance_by_size(partitions, model, x)
model = GPipe(model, balance, chunks=4, checkpoint="always")
# config loss functions
dice_loss_func = DiceLoss(softmax=True, reduction="none")
# use the same pipeline and loss in
# AnatomyNet: Deep learning for fast and fully automated whole‐volume segmentation of head and neck anatomy,
# Medical Physics, 2018.
focal_loss_func = FocalLoss(reduction="none")
if pretrain:
print(f"loading from {pretrain}.")
pretrained_dict = torch.load(pretrain)["weight"]
model_dict = model.state_dict()
pretrained_dict = {k: v for k, v in pretrained_dict.items() if k in model_dict}
model_dict.update(pretrained_dict)
model.load_state_dict(pretrained_dict)
b_time = time.time()
best_val_loss = [0] * (N_CLASSES - 1) # foreground
for epoch in range(ep):
model.train()
trainloss = 0
for b_idx, data_dict in enumerate(train_dataloader):
x_train = data_dict["image"]
y_train = data_dict["label"]
flagvec = data_dict["with_complete_groundtruth"]
x_train = torch.autograd.Variable(x_train.cuda())
y_train = torch.autograd.Variable(y_train.cuda().float())
optimizer.zero_grad()
o = model(x_train).to(0, non_blocking=True).float()
loss = (dice_loss_func(o, y_train.to(o)) * flagvec.to(o) * lossweight.to(o)).mean()
loss += 0.5 * (focal_loss_func(o, y_train.to(o)) * flagvec.to(o) * lossweight.to(o)).mean()
loss.backward()
optimizer.step()
trainloss += loss.item()
if b_idx % 20 == 0:
print(f"Train Epoch: {epoch} [{b_idx}/{len(train_dataloader)}] \tLoss: {loss.item()}")
print(f"epoch {epoch} TRAIN loss {trainloss / len(train_dataloader)}")
if epoch % 10 == 0:
model.eval()
# check validation dice
val_loss = [0] * (N_CLASSES - 1)
n_val = [0] * (N_CLASSES - 1)
for data_dict in val_dataloader:
x_val = data_dict["image"]
y_val = data_dict["label"]
with torch.no_grad():
x_val = torch.autograd.Variable(x_val.cuda())
o = model(x_val).to(0, non_blocking=True)
loss = compute_meandice(o, y_val.to(o), mutually_exclusive=True, include_background=False)
val_loss = [l.item() + tl if l == l else tl for l, tl in zip(loss[0], val_loss)]
n_val = [n + 1 if l == l else n for l, n in zip(loss[0], n_val)]
val_loss = [l / n for l, n in zip(val_loss, n_val)]
print("validation scores %.4f, %.4f, %.4f, %.4f, %.4f, %.4f, %.4f, %.4f, %.4f" % tuple(val_loss))
for c in range(1, 10):
if best_val_loss[c - 1] < val_loss[c - 1]:
best_val_loss[c - 1] = val_loss[c - 1]
state = {"epoch": epoch, "weight": model.state_dict(), "score_" + str(c): best_val_loss[c - 1]}
torch.save(state, f"{model_name}" + str(c))
print("best validation scores %.4f, %.4f, %.4f, %.4f, %.4f, %.4f, %.4f, %.4f, %.4f" % tuple(best_val_loss))
print("total time", time.time() - b_time)
if __name__ == "__main__":
parser = ArgumentParser()
parser.add_argument("--n_feat", type=int, default=32, dest="n_feat")
parser.add_argument("--crop_size", type=str, default="-1,-1,-1", dest="crop_size")
parser.add_argument("--bs", type=int, default=1, dest="bs") # batch size
parser.add_argument("--ep", type=int, default=150, dest="ep") # number of epochs
parser.add_argument("--lr", type=float, default=5e-4, dest="lr") # learning rate
parser.add_argument("--optimizer", type=str, default="rmsprop", dest="optimizer") # type of optimizer
parser.add_argument("--pretrain", type=str, default=None, dest="pretrain")
args = parser.parse_args()
input_dict = vars(args)
print(input_dict)
train(**input_dict)