test during training default to FP16

pull/1/head
Glenn Jocher 5 years ago
parent bd3e389130
commit a1748a8d6e

@ -23,6 +23,7 @@ def test(data,
verbose=False):
# Initialize/load model and set device
if model is None:
training = False
device = torch_utils.select_device(opt.device, batch_size=batch_size)
half = device.type != 'cpu' # half precision only supported on CUDA
@ -42,11 +43,12 @@ def test(data,
if device.type != 'cpu' and torch.cuda.device_count() > 1:
model = nn.DataParallel(model)
training = False
else: # called by train.py
device = next(model.parameters()).device # get model device
half = False
training = True
device = next(model.parameters()).device # get model device
half = device.type != 'cpu' # half precision only supported on CUDA
if half:
model.half() # to FP16
# Configure
model.eval()

Loading…
Cancel
Save