瀏覽代碼

Resume with custom anchors fix (#2361)

* Resume with custom anchors fix

* Update train.py
5.0
Glenn Jocher GitHub 3 年之前
父節點
當前提交
e931b9da33
沒有發現已知的金鑰在資料庫的簽署中 GPG Key ID: 4AEE18F83AFDEB23
共有 1 個文件被更改,包括 3 次插入4 次删除
  1. +3
    -4
      train.py

+ 3
- 4
train.py 查看文件

@@ -75,10 +75,8 @@ def train(hyp, opt, device, tb_writer=None, wandb=None):
with torch_distributed_zero_first(rank):
attempt_download(weights) # download if not found locally
ckpt = torch.load(weights, map_location=device) # load checkpoint
if hyp.get('anchors'):
ckpt['model'].yaml['anchors'] = round(hyp['anchors']) # force autoanchor
model = Model(opt.cfg or ckpt['model'].yaml, ch=3, nc=nc).to(device) # create
exclude = ['anchor'] if opt.cfg or hyp.get('anchors') else [] # exclude keys
model = Model(opt.cfg or ckpt['model'].yaml, ch=3, nc=nc, anchors=hyp.get('anchors')).to(device) # create
exclude = ['anchor'] if (opt.cfg or hyp.get('anchors')) and not opt.resume else [] # exclude keys
state_dict = ckpt['model'].float().state_dict() # to FP32
state_dict = intersect_dicts(state_dict, model.state_dict(), exclude=exclude) # intersect
model.load_state_dict(state_dict, strict=False) # load
@@ -216,6 +214,7 @@ def train(hyp, opt, device, tb_writer=None, wandb=None):
# Anchors
if not opt.noautoanchor:
check_anchors(dataset, model=model, thr=hyp['anchor_t'], imgsz=imgsz)
model.half().float() # pre-reduce anchor precision

# Model parameters
hyp['box'] *= 3. / nl # scale to layers

Loading…
取消
儲存