|
@@ -33,7 +33,7 @@ def train_one_epoch(model, optimizer, data_loader, device, epoch, print_freq, sc
|
|
targets = [{k: v.to(device) if isinstance(v, torch.Tensor) else v for k, v in t.items()} for t in targets]
|
|
targets = [{k: v.to(device) if isinstance(v, torch.Tensor) else v for k, v in t.items()} for t in targets]
|
|
with torch.cuda.amp.autocast(enabled=scaler is not None):
|
|
with torch.cuda.amp.autocast(enabled=scaler is not None):
|
|
loss_dict = model(images, targets)
|
|
loss_dict = model(images, targets)
|
|
- print(f'loss_dict:{loss_dict}')
|
|
|
|
|
|
+ # print(f'loss_dict:{loss_dict}')
|
|
losses = sum(loss for loss in loss_dict.values())
|
|
losses = sum(loss for loss in loss_dict.values())
|
|
|
|
|
|
# reduce losses over all GPUs for logging purposes
|
|
# reduce losses over all GPUs for logging purposes
|