|
|
|
@ -645,26 +645,18 @@ def non_max_suppression(prediction, conf_thres=0.1, iou_thres=0.6, merge=False,
|
|
|
|
|
return output
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def strip_optimizer(f='weights/best.pt'): # from utils.utils import *; strip_optimizer()
|
|
|
|
|
# Strip optimizer from *.pt files for lighter files (reduced by 1/2 size)
|
|
|
|
|
x = torch.load(f, map_location=torch.device('cpu'))
|
|
|
|
|
x['optimizer'] = None
|
|
|
|
|
x['model'].half() # to FP16
|
|
|
|
|
torch.save(x, f)
|
|
|
|
|
print('Optimizer stripped from %s, %.1fMB' % (f, os.path.getsize(f) / 1E6))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def create_pretrained(f='weights/best.pt', s='weights/pretrained.pt'): # from utils.utils import *; create_pretrained()
|
|
|
|
|
# create pretrained checkpoint 's' from 'f' (create_pretrained(x, x) for x in glob.glob('./*.pt'))
|
|
|
|
|
def strip_optimizer(f='weights/best.pt', s=''): # from utils.utils import *; strip_optimizer()
|
|
|
|
|
# Strip optimizer from 'f' to finalize training, optionally save as 's'
|
|
|
|
|
x = torch.load(f, map_location=torch.device('cpu'))
|
|
|
|
|
x['optimizer'] = None
|
|
|
|
|
x['training_results'] = None
|
|
|
|
|
x['epoch'] = -1
|
|
|
|
|
x['model'].half() # to FP16
|
|
|
|
|
for p in x['model'].parameters():
|
|
|
|
|
p.requires_grad = True
|
|
|
|
|
torch.save(x, s)
|
|
|
|
|
print('%s saved as pretrained checkpoint %s, %.1fMB' % (f, s, os.path.getsize(s) / 1E6))
|
|
|
|
|
p.requires_grad = False
|
|
|
|
|
torch.save(x, s or f)
|
|
|
|
|
mb = os.path.getsize(s or f) / 1E6 # filesize
|
|
|
|
|
print('Optimizer stripped from %s,%s %.1fMB' % (f, (' saved as %s,' % s) if s else '', mb))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def coco_class_count(path='../coco/labels/train2014/'):
|
|
|
|
|